repo_name
stringlengths
5
92
path
stringlengths
4
232
copies
stringclasses
19 values
size
stringlengths
4
7
content
stringlengths
721
1.04M
license
stringclasses
15 values
hash
int64
-9,223,277,421,539,062,000
9,223,102,107B
line_mean
float64
6.51
99.9
line_max
int64
15
997
alpha_frac
float64
0.25
0.97
autogenerated
bool
1 class
agabert/zeus
stages/keystone/fabfile.py
1
5050
import os from zeus.config import ConfigManager from zeus.common import FabricManager from zeus.common import PasswordManager from zeus.configmanagement import ConfigEditor from zeus.ubuntu import RepoManager from zeus.services import ServiceControl from fabric.api import parallel, roles, run, env metadata = ConfigManager(os.environ["CONFIGFILE"]) passwords = PasswordManager(os.environ["PASSWORDCACHE"]).passwords FabricManager.setup(metadata.roles_ports) # # http://docs.openstack.org/mitaka/install-guide-ubuntu/keystone-install.html # @parallel @roles('openstack_keystone') def keystone(): run(""" echo "manual" > /etc/init/keystone.override """) RepoManager.install("keystone") run(""" systemctl stop keystone; echo systemctl disable keystone; echo """) RepoManager.install("apache2") RepoManager.install("libapache2-mod-wsgi") ConfigEditor.setKey( "/etc/keystone/keystone.conf", "DEFAULT", "admin_token", passwords["ADMIN_TOKEN"]) ConfigEditor.setKey( "/etc/keystone/keystone.conf", "database", "connection", "mysql+pymysql://keystone:%s@%s/keystone" % ( passwords["KEYSTONE_DBPASS"], metadata.servers[metadata.roles['openstack_mysql'][0]]['ip'])) ConfigEditor.setKey( "/etc/keystone/keystone.conf", "token", "provider", "fernet") run(""" su -s /bin/sh -c "keystone-manage db_sync" keystone """) run(""" keystone-manage fernet_setup --keystone-user keystone --keystone-group keystone """) run(""" uudecode -o /etc/apache2/apache2.conf <<EOF begin-base64-encoded 644 YXBhY2hlMi5jb25m U2VydmVyTmFtZSBYWFhYWFhYWFgKTXV0ZXggZmlsZToke0FQQUNIRV9MT0NL X0RJUn0gZGVmYXVsdApQaWRGaWxlICR7QVBBQ0hFX1BJRF9GSUxFfQpUaW1l b3V0IDMwMApLZWVwQWxpdmUgT24KTWF4S2VlcEFsaXZlUmVxdWVzdHMgMTAw CktlZXBBbGl2ZVRpbWVvdXQgNQpVc2VyICR7QVBBQ0hFX1JVTl9VU0VSfQpH cm91cCAke0FQQUNIRV9SVU5fR1JPVVB9Ckhvc3RuYW1lTG9va3VwcyBPZmYK RXJyb3JMb2cgJHtBUEFDSEVfTE9HX0RJUn0vZXJyb3IubG9nCkxvZ0xldmVs IHdhcm4KSW5jbHVkZU9wdGlvbmFsIG1vZHMtZW5hYmxlZC8qLmxvYWQKSW5j bHVkZU9wdGlvbmFsIG1vZHMtZW5hYmxlZC8qLmNvbmYKSW5jbHVkZSBwb3J0 cy5jb25mCjxEaXJlY3RvcnkgLz4KCU9wdGlvbnMgRm9sbG93U3ltTGlua3MK CUFsbG93T3ZlcnJpZGUgTm9uZQoJUmVxdWlyZSBhbGwgZGVuaWVkCjwvRGly ZWN0b3J5Pgo8RGlyZWN0b3J5IC91c3Ivc2hhcmU+CglBbGxvd092ZXJyaWRl IE5vbmUKCVJlcXVpcmUgYWxsIGdyYW50ZWQKPC9EaXJlY3Rvcnk+CjxEaXJl Y3RvcnkgL3Zhci93d3cvPgoJT3B0aW9ucyBJbmRleGVzIEZvbGxvd1N5bUxp bmtzCglBbGxvd092ZXJyaWRlIE5vbmUKCVJlcXVpcmUgYWxsIGdyYW50ZWQK PC9EaXJlY3Rvcnk+CkFjY2Vzc0ZpbGVOYW1lIC5odGFjY2Vzcwo8RmlsZXNN YXRjaCAiXlwuaHQiPgoJUmVxdWlyZSBhbGwgZGVuaWVkCjwvRmlsZXNNYXRj aD4KTG9nRm9ybWF0ICIldjolcCAlaCAlbCAldSAldCBcIiVyXCIgJT5zICVP IFwiJXtSZWZlcmVyfWlcIiBcIiV7VXNlci1BZ2VudH1pXCIiIHZob3N0X2Nv bWJpbmVkCkxvZ0Zvcm1hdCAiJWggJWwgJXUgJXQgXCIlclwiICU+cyAlTyBc IiV7UmVmZXJlcn1pXCIgXCIle1VzZXItQWdlbnR9aVwiIiBjb21iaW5lZApM b2dGb3JtYXQgIiVoICVsICV1ICV0IFwiJXJcIiAlPnMgJU8iIGNvbW1vbgpM b2dGb3JtYXQgIiV7UmVmZXJlcn1pIC0+ICVVIiByZWZlcmVyCkxvZ0Zvcm1h dCAiJXtVc2VyLWFnZW50fWkiIGFnZW50CkluY2x1ZGVPcHRpb25hbCBjb25m LWVuYWJsZWQvKi5jb25mCkluY2x1ZGVPcHRpb25hbCBzaXRlcy1lbmFibGVk LyouY29uZgo= ==== EOF """) this = env.host_string.split(":")[0] run(""" sed -i 's,XXXXXXXXX,%s,g;' /etc/apache2/apache2.conf """ % this) run(""" uudecode -o /etc/apache2/sites-available/wsgi-keystone.conf<<EOF begin-base64-encoded 644 d3NnaS1rZXlzdG9uZS5jb25m Ckxpc3RlbiA1MDAwCkxpc3RlbiAzNTM1NwoKPFZpcnR1YWxIb3N0ICo6NTAw MD4KICAgIFdTR0lEYWVtb25Qcm9jZXNzIGtleXN0b25lLXB1YmxpYyBwcm9j ZXNzZXM9NSB0aHJlYWRzPTEgdXNlcj1rZXlzdG9uZSBncm91cD1rZXlzdG9u ZSBkaXNwbGF5LW5hbWU9JXtHUk9VUH0KICAgIFdTR0lQcm9jZXNzR3JvdXAg a2V5c3RvbmUtcHVibGljCiAgICBXU0dJU2NyaXB0QWxpYXMgLyAvdXNyL2Jp bi9rZXlzdG9uZS13c2dpLXB1YmxpYwogICAgV1NHSUFwcGxpY2F0aW9uR3Jv dXAgJXtHTE9CQUx9CiAgICBXU0dJUGFzc0F1dGhvcml6YXRpb24gT24KICAg IEVycm9yTG9nRm9ybWF0ICIle2N1fXQgJU0iCiAgICBFcnJvckxvZyAvdmFy L2xvZy9hcGFjaGUyL2tleXN0b25lLmxvZwogICAgQ3VzdG9tTG9nIC92YXIv bG9nL2FwYWNoZTIva2V5c3RvbmVfYWNjZXNzLmxvZyBjb21iaW5lZAoKICAg IDxEaXJlY3RvcnkgL3Vzci9iaW4+CiAgICAgICAgUmVxdWlyZSBhbGwgZ3Jh bnRlZAogICAgPC9EaXJlY3Rvcnk+CjwvVmlydHVhbEhvc3Q+Cgo8VmlydHVh bEhvc3QgKjozNTM1Nz4KICAgIFdTR0lEYWVtb25Qcm9jZXNzIGtleXN0b25l LWFkbWluIHByb2Nlc3Nlcz01IHRocmVhZHM9MSB1c2VyPWtleXN0b25lIGdy b3VwPWtleXN0b25lIGRpc3BsYXktbmFtZT0le0dST1VQfQogICAgV1NHSVBy b2Nlc3NHcm91cCBrZXlzdG9uZS1hZG1pbgogICAgV1NHSVNjcmlwdEFsaWFz IC8gL3Vzci9iaW4va2V5c3RvbmUtd3NnaS1hZG1pbgogICAgV1NHSUFwcGxp Y2F0aW9uR3JvdXAgJXtHTE9CQUx9CiAgICBXU0dJUGFzc0F1dGhvcml6YXRp b24gT24KICAgIEVycm9yTG9nRm9ybWF0ICIle2N1fXQgJU0iCiAgICBFcnJv ckxvZyAvdmFyL2xvZy9hcGFjaGUyL2tleXN0b25lLmxvZwogICAgQ3VzdG9t TG9nIC92YXIvbG9nL2FwYWNoZTIva2V5c3RvbmVfYWNjZXNzLmxvZyBjb21i aW5lZAoKICAgIDxEaXJlY3RvcnkgL3Vzci9iaW4+CiAgICAgICAgUmVxdWly ZSBhbGwgZ3JhbnRlZAogICAgPC9EaXJlY3Rvcnk+CjwvVmlydHVhbEhvc3Q+ Cgo= ==== EOF a2enmod wsgi ln -sf /etc/apache2/sites-available/wsgi-keystone.conf /etc/apache2/sites-enabled rm -f /var/lib/keystone/keystone.db """) ServiceControl.relaunch("apache2") ServiceControl.check("wsgi:keystone")
apache-2.0
8,810,368,617,381,399,000
33.827586
81
0.852475
false
GregorCH/ipet
ipet/parsing/StatisticReader_CustomReader.py
1
6339
""" The MIT License (MIT) Copyright (c) 2018 Zuse Institute Berlin, www.zib.de Permissions are granted as stated in the license file you have obtained with this software. If you find the library useful for your purpose, please refer to README.md for how to cite IPET. @author: Gregor Hendel """ from .StatisticReader import StatisticReader import re import builtins import logging from ipet import misc from ipet.concepts.IPETNode import IpetNode logger = logging.getLogger(__name__) class CustomReader(StatisticReader): """ Reader to be initialised interactively through IPET or from an interactive python shell """ name = 'CustomReader' regexp = 'Custom' datakey = 'Custom' data = None METHOD_FIRST = 1 METHOD_LAST = 2 METHOD_SUM = 3 METHOD_MIN = 4 METHOD_MAX = 5 METHOD_COUNT = 6 str2method = { "first" : METHOD_FIRST, "last" : METHOD_LAST, "sum" : METHOD_SUM, "min" : METHOD_MIN, "max" : METHOD_MAX, "count" : METHOD_COUNT } requiredoptions = { "datatype" : ["float", "int"], "method" : list(str2method.keys()) } def __init__(self, name = None, regpattern = None, datakey = None, index = 0, datatype = "float", method = "last", active = True): """ constructor of a custom reader to parse additional simple solver output from log file context Parameters: ----------- name : a name to distinguish this reader from the others regpattern : A string or regular expression pattern to detect lines from which output should be read datakey : The data key under which the parsed datum gets stored for every problem index : The zero-based index of the number in the specified line (only numbers count) datatype : choose 'int' or 'float' method : how to treat multiple occurrences of this data within one problem; 'count' occurrences or parse 'first', 'last', 'sum', 'min' or 'max' """ IpetNode.__init__(self, active) if regpattern is None: raise ValueError("Error: No 'regpattern' specified for reader with name %s" % str(name)) if name in [None, ""]: self.name = datakey + "Reader" self.username = False else: self.name = name self.username = True self.set_datakey(datakey) self.set_index(index) self.regpattern = regpattern self.set_regpattern(regpattern) self.method = method self.methodint = self.METHOD_LAST self.set_method(method) self.set_datatype(datatype) def getEditableAttributes(self): return ['name', 'regpattern', 'datakey', 'index', 'datatype', 'method'] + IpetNode.getEditableAttributes(self) def getRequiredOptionsByAttribute(self, attr): return self.requiredoptions.get(attr, IpetNode.getRequiredOptionsByAttribute(self, attr)) def extractStatistic(self, line): if self.regexp.search(line): logging.debug("Custom Reader {} found match in line \n{}".format(self.name, line.strip())) logging.debug("Numerical expression matches: {}".format(", ".join(misc.numericExpression.findall(line)))) previousdata = self.testrun.getCurrentProblemData(self.datakey) if self.methodint == CustomReader.METHOD_COUNT: if previousdata is None: self.addData(self.datakey, 1) else: self.addData(self.datakey, previousdata + 1) return try: data = misc.getNumberAtIndex(line, self.index) data = self.datatypemethod(data) if self.methodint == CustomReader.METHOD_FIRST: if previousdata is None: self.addData(self.datakey, data) elif self.methodint == CustomReader.METHOD_LAST: self.addData(self.datakey, data) elif self.methodint == CustomReader.METHOD_SUM: if previousdata is None: previousdata = 0 self.addData(self.datakey, data + previousdata) elif self.methodint == CustomReader.METHOD_MIN: if previousdata is None: self.addData(self.datakey, data) elif data < previousdata: self.addData(self.datakey, data) elif self.methodint == CustomReader.METHOD_MAX: if previousdata is None: self.addData(self.datakey, data) elif data > previousdata: self.addData(self.datakey, data) except: logger.warn("Reader %s could not retrieve data at index %d from matching line '%s'", self.getName(), self.index, line) pass return None def setDataType(self, sometype): """ recognizes data types (e.g., 'float' or 'int') and sets reader data type to this value """ try: self.datatypemethod = getattr(builtins, sometype) self.datatype = sometype except: logger.debug("Error: Could not recognize data type %s, using float" % sometype) self.datatypemethod = float self.datatype = 'float' def set_datatype(self, datatype): self.setDataType(datatype) def set_method(self, method): self.methodint = self.str2method.get(method, self.methodint) self.method = method def set_regpattern(self, regpattern): self.regexp = re.compile(regpattern) self.regpattern = regpattern def set_name(self, name): if name == self.getName(): return if name in ["", None]: self.name = self.datakey + "Reader" self.username = False else: self.name = name self.username = True def set_datakey(self, datakey): self.datakey = datakey if not self.username: self.name = self.datakey + "Reader" def set_index(self, index): self.index = int(index)
mit
-7,036,275,254,161,661,000
31.84456
151
0.581953
false
jackfirth/flask-negotiate
setup.py
1
1312
""" flask-negotiate2 =============== Content negotiation utility for Flask apps. Fork of Flask-Negotiate by Matt Wright (github.com/mattupstate) Resources --------- - `Documentation <http://packages.python.org/flask-negotiate2/>`_ - `Issue Tracker <http://github.com/jackfirth/flask-negotiate/issues>`_ - `Code <http://github.com/jackfirth/flask-negotiate/>`_ """ from setuptools import setup setup( name='flask-negotiate2', version='0.2.0', url='https://github.com/jackfirth/flask-negotiate', license='MIT', author='Matthew Wright', author_email='[email protected]', description='Content negotiation utility for Flask apps', long_description=__doc__, py_modules=['flask_negotiate2'], zip_safe=False, include_package_data=True, platforms='any', install_requires=['flask'], test_suite='nose.collector', tests_require=['nose'], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
mit
-3,190,932,591,787,335,000
28.155556
71
0.645579
false
NoneGroupTeam/Let-s-Go
webapp/app/views.py
1
18587
from django.contrib.auth import authenticate, logout, login from django.contrib.auth import get_user_model from django.http import HttpResponse from app.models import Label, Place, Guide, Question, Answer from hashlib import md5 import json import re LOGIN_OK_CODE = 200 LOGIN_OK = 'Login success' LOGOUT_OK_CODE = 201 LOGOUT_OK = 'Logout success' REG_OK_CODE = 202 REG_OK = 'Regist success' QUERY_OK_CODE = 203 QUERY_OK = '' ADD_OK_CODE = 203 ADD_OK = 'Add success' GET_OK_CODE = 204 GET_OK = '' HAD_LOGIN_CODE = 301 HAD_LOGIN = 'Had logined' NOT_LOGIN_CODE = 301 NOT_LOGIN = 'Not login' NOT_ACTIVE_CODE = 401 NOT_ACTIVE = 'User Not Active' NOT_MATCH_CODE = 402 NOT_MATCH = 'Username and Password not match' DATE_ERR_CODE = 411 DATE_ERR = 'Datetime is not allow' GENDER_ERR_CODE = 412 GENDER_ERR = 'Gender is not allow' PHONE_ERR_CODE = 413 PHONE_ERR = 'Phone num is not allow' EMAIL_ERR_CODE = 414 EMAIL_ERR = 'Email is not allow' PHONE_EX_CODE = 421 PHONE_EX = 'Phone has already regist' EMAIL_EX_CODE = 422 EMAIL_EX = 'Email has already regist' UNAME_EX_CODE = 423 UNAME_EX = 'Username has already regist' NAME_EX_CODE = 424 NAME_EX = 'This name is already exists' KEY_ERR_CODE = 425 KEY_ERR = 'The Key Error' ID_ERR_CODE = 426 ID_ERR = 'The ID Error' TITLE_ERR_CODE = 427 TITLE_ERR = 'The Title Error' PLACE_ERR_CODE = 428 PLACE_ERR = 'The Place Error' LABEL_ERR_CODE = 429 LABEL_ERR = 'The Label Error' NAME_ERR_CODE = 430 NAME_ERR = 'Name Error' NAME_NEX_CODE = 431 NAME_NEX = 'Name Not exists' INVALIED_CODE = 501 INVALIED = 'Not support this method' UN_ERROR_CODE = 502 UN_ERROR = 'Something error' def index(request): return HttpResponse("hello.") def test(request): return HttpResponse("test ok") def JSON(**kwargs): return json.dumps(kwargs) def user_logout(request): if request.user.is_authenticated(): logout(request) data = JSON(code=LOGOUT_OK_CODE, status=True, message=LOGOUT_OK) else: data = JSON(code=NOT_LOGIN_CODE, status=True, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def user_login(request): if request.user.is_authenticated(): data = JSON(code=HAD_LOGIN_CODE, status=True, message=HAD_LOGIN) return HttpResponse(data, content_type="application/json") if request.method == 'POST': username = request.POST.get('username') password = request.POST.get('password') user = authenticate(username=username, password=password) if user is not None: if user.is_active: message = JSON(user_id=user.id, username=user.username) data = JSON(code=LOGIN_OK_CODE, status=True, message=message) login(request, user) else: data = JSON(code=NOT_ACTIVE_CODE, status=False, message=NOT_ACTIVE) else: data = JSON(code=NOT_MATCH_CODE, status=False, message=NOT_MATCH) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) return HttpResponse(data, content_type="application/json") def user_register(request): if request.user.is_authenticated(): data = JSON(code=HAD_LOGIN_CODE, status=False, message=HAD_LOGIN) elif not request.method == 'POST': data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: username = request.POST.get('username') password = request.POST.get('password') email = request.POST.get('email') phone = request.POST.get('phone') gender = request.POST.get('gender') birthday = request.POST.get('birthday') # check format if re.match(r'(\d{4}([-/\.])\d{2}\2\d{2})', birthday) is None: data = JSON(code=DATE_ERR_CODE, status=False, message=DATE_ERR) elif gender not in {'1', '0'}: data = JSON(code=GENDER_ERR_CODE, status=False, message=GENDER_ERR) elif re.match(r'(\+\d{1,3})?1\d{10}', phone) is None: data = JSON(code=PHONE_ERR_CODE, status=False, message=PHONE_ERR) elif re.match(r'[^@\s]+@([^@\s]+\.)+[^@\s]+', email) is None: data = JSON(code=EMAIL_ERR_CODE, status=False, message=EMAIL_ERR) # database search else: all_user = get_user_model().objects if all_user.filter(phone=phone).count() != 0: data = JSON(CODE=PHONE_EX_CODE, status=False, message=PHONE_EX) elif all_user.filter(email=email).count() != 0: data = JSON(CODE=EMAIL_EX_CODE, status=False, message=EMAIL_EX) elif all_user.filter(username=username).count() != 0: data = JSON(CODE=UNAME_EX_CODE, status=False, message=UNAME_EX) else: app_user = get_user_model() try: birthday = birthday.replace('.', '-').replace('/', '-') user = app_user.objects.create_user(username=username, password=password, email=email, phone=phone, gender=gender, birthday=birthday) message = JSON(user_id=user.id, username=user.username) data = JSON(code=REG_OK_CODE, status=True, message=message) except Exception as e: print(e) data = JSON(code=UN_ERROR_CODE, status=False, message=UN_ERROR) return HttpResponse(data, content_type="application/json") def guide_add(request): if request.user.is_authenticated(): if request.method == 'POST': title = request.POST.get('title') content = request.POST.get('content') place = request.POST.get('place') label = request.POST.getlist('label[]') start_time = request.POST.get('start_time') end_time = request.POST.get('end_time') if len(title) == 0: data = JSON(code=TITLE_ERR_CODE, status=False, message=TITLE_ERR) elif len(place) == 0: data = JSON(code=PLACE_ERR_CODE, status=False, message=PLACE_ERR) elif re.match(r'(\d{4}([-/\.])\d{2}\2\d{2})', start_time) is None: data = JSON(code=DATE_ERR_CODE, status=False, message=DATE_ERR) elif re.match(r'(\d{4}([-/\.])\d{2}\2\d{2})', end_time) is None: data = JSON(code=DATE_ERR_CODE, status=False, message=DATE_ERR) elif start_time > end_time: data = JSON(code=DATE_ERR_CODE, status=False, message=DATE_ERR) elif not Place.objects.filter(id=place): data = JSON(code=PLACE_ERR_CODE, status=False, message=PLACE_ERR) else: label = Label.objects.filter(id__in=label) a = Guide(name=title, user=request.user, place=Place.objects.get(id=place), content=content, start_time=start_time, end_time=end_time) a.save() a.label.add(*label) data = JSON(code=ADD_OK_CODE, status=True, message=ADD_OK) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def guide_id(request, _id): if request.user.is_authenticated(): try: guide = Guide.objects.filter(id=_id)[0] labels = [] for l in guide.label.all(): labels.append(l.name) submit = str(guide.submit.strftime('%Y-%m-%d %H:%M:%S')) result = {'title': guide.name, 'username': guide.user.username, 'place': guide.place.name, 'labels': labels, 'start_time': str(guide.start_time), 'end_time': str(guide.end_time), 'content': guide.content, 'submit': submit, 'pageview': guide.pageview} guide.pageview += 1 guide.save() data = JSON(code=GET_OK_CODE, status=True, message=result) except IndexError: data = JSON(code=ID_ERR_CODE, status=False, message=ID_ERR) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def guide_list(request): if request.user.is_authenticated(): if request.method == 'POST': start = int(request.POST.get('start')) offset = int(request.POST.get('offset')) try: ans = Guide.objects.order_by('-id')[start:start + offset] except IndexError: ans = [] result = [] for i in ans: labels = [] for l in i.label.all(): labels.append(l.name) m = md5() m.update(i.user.email.encode()) img = 'http://gravatar.eqoe.cn/avatar/%s?size=48&default=identicon&rating=pg' % (m.hexdigest()) _ = {'id': i.id, 'username': i.user.username, 'title': i.name, 'place': i.place.name, 'pageview': i.pageview, 'labels': labels, 'img': img} result.append(_) data = JSON(code=QUERY_OK_CODE, status=True, message=result) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def question_add(request): if request.user.is_authenticated(): if request.method == 'POST': title = request.POST.get('title') content = request.POST.get('content') place = request.POST.get('place') label = request.POST.getlist('label[]') if len(title) == 0: data = JSON(code=TITLE_ERR_CODE, status=False, message=TITLE_ERR) elif len(place) == 0: data = JSON(code=PLACE_ERR_CODE, status=False, message=PLACE_ERR) elif not Place.objects.filter(id=place): data = JSON(code=PLACE_ERR_CODE, status=False, message=PLACE_ERR) else: label = Label.objects.filter(id__in=label) a = Question(title=title, user=request.user, place=Place.objects.get(id=place), content=content) a.save() a.label.add(*label) data = JSON(code=ADD_OK_CODE, status=True, message=ADD_OK) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def question_id(request, _id): if request.user.is_authenticated(): try: question = Question.objects.filter(id=_id)[0] labels = [] for l in question.label.all(): labels.append(l.name) answers = [] for i in Answer.objects.filter(question=question).order_by('-submit'): m = md5() m.update(i.user.email.encode()) img = 'http://gravatar.eqoe.cn/avatar/%s?size=48&default=identicon&rating=pg' % (m.hexdigest()) _submit = str(i.submit.strftime('%Y-%m-%d %H:%M:%S')) _ = {'id': i.id, 'username': i.user.username, 'img': img, 'content': i.content, 'submit': _submit} answers.append(_) submit = str(question.submit.strftime('%Y-%m-%d %H:%M:%S')) result = {'title': question.title, 'username': question.user.username, 'place': question.place.name, 'labels': labels, 'content': question.content, 'submit': submit, 'answer': answers} data = JSON(code=GET_OK_CODE, status=True, message=result) except IndexError: data = JSON(code=ID_ERR_CODE, status=False, message=ID_ERR) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def question_comment(request, _id): if request.user.is_authenticated(): if request.method == 'POST': content = request.POST.get('content') try: question = Question.objects.filter(id=_id)[0] answer = Answer(user=request.user, question=question, content=content) answer.save() data = JSON(code=ADD_OK_CODE, status=True, message=ADD_OK) except IndexError: data = JSON(code=ID_ERR_CODE, status=False, message=ID_ERR) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def question_list(request): if request.user.is_authenticated(): if request.method == 'POST': start = int(request.POST.get('start')) offset = int(request.POST.get('offset')) try: ans = Question.objects.order_by('-id')[start:start + offset] except IndexError: ans = [] result = [] for i in ans: labels = [] for l in i.label.all(): labels.append(l.name) m = md5() m.update(i.user.email.encode()) ans_count = len(Answer.objects.filter(question=i)) img = 'http://gravatar.eqoe.cn/avatar/%s?size=48&default=identicon&rating=pg' % (m.hexdigest()) _ = {'id': i.id, 'username': i.user.username, 'title': i.title, 'place': i.place.name, 'answer': ans_count, 'labels': labels, 'img': img} result.append(_) data = JSON(code=QUERY_OK_CODE, status=True, message=result) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def __id(request, _id, model): if request.user.is_authenticated(): try: ans = model.objects.filter(id=_id)[0].name data = JSON(code=QUERY_OK_CODE, status=True, message=ans) except IndexError: data = JSON(code=ID_ERR_CODE, status=False, message=ID_ERR) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def label_id(request, _id): return __id(request, _id, Label) def place_id(request, _id): return __id(request, _id, Place) def __list(request, model): if request.user.is_authenticated(): ans = list(model.objects.values('id', 'name')) data = JSON(code=QUERY_OK_CODE, status=True, message=ans) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def place_list(request): return __list(request, Place) def label_list(request): return __list(request, Label) def user_add_place(request): if request.user.is_authenticated(): if request.method == 'POST': if 'name' in request.POST: name = request.POST.get('name') if len(name) == 0: data = data = JSON(code=NAME_ERR_CODE, status=True, message=NAME_ERR) elif not Place.objects.filter(name=name): data = JSON(code=NAME_NEX_CODE, status=False, message=NAME_NEX) else: request.user.place.add(Place.objects.get(name=name)) data = JSON(code=ADD_OK_CODE, status=True, message=ADD_OK) else: data = JSON(code=KEY_ERR_CODE, status=False, message=KEY_ERR) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def __add(request, model): if request.user.is_authenticated(): if request.method == 'POST': if 'name' in request.POST: name = request.POST.get('name') if len(name) == 0: data = data = JSON(code=NAME_ERR_CODE, status=True, message=NAME_ERR) elif model.objects.filter(name=name): data = JSON(code=NAME_EX_CODE, status=False, message=NAME_EX) else: add = model(name=name) add.save() data = JSON(code=ADD_OK_CODE, status=True, message=ADD_OK) else: data = JSON(code=KEY_ERR_CODE, status=False, message=KEY_ERR) else: data = JSON(code=INVALIED_CODE, status=False, message=INVALIED) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json") def label_add(request): return __add(request, Label) def place_add(request): return __add(request, Place) def user_info(request): if request.user.is_authenticated(): I = request.user places = [] for l in I.place.all(): places.append(l.name) result = {'username': I.username, 'id': I.id, 'places': places, 'birthday': str(I.birthday), 'gender': I.gender} data = JSON(code=GET_OK_CODE, status=True, message=result) else: data = JSON(code=NOT_LOGIN_CODE, status=False, message=NOT_LOGIN) return HttpResponse(data, content_type="application/json")
gpl-3.0
-4,711,005,114,202,626,000
38.63113
111
0.555334
false
AutorestCI/azure-sdk-for-python
azure-mgmt-servicebus/azure/mgmt/servicebus/operations/operations.py
1
3676
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- import uuid from msrest.pipeline import ClientRawResponse from .. import models class Operations(object): """Operations operations. :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An objec model deserializer. :ivar api_version: Client API version. Constant value: "2017-04-01". """ models = models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self.api_version = "2017-04-01" self.config = config def list( self, custom_headers=None, raw=False, **operation_config): """Lists all of the available ServiceBus REST API operations. :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: An iterator like instance of Operation :rtype: ~azure.mgmt.servicebus.models.OperationPaged[~azure.mgmt.servicebus.models.Operation] :raises: :class:`ErrorResponseException<azure.mgmt.servicebus.models.ErrorResponseException>` """ def internal_paging(next_link=None, raw=False): if not next_link: # Construct URL url = '/providers/Microsoft.ServiceBus/operations' # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') else: url = next_link query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send( request, header_parameters, **operation_config) if response.status_code not in [200]: raise models.ErrorResponseException(self._deserialize, response) return response # Deserialize response deserialized = models.OperationPaged(internal_paging, self._deserialize.dependencies) if raw: header_dict = {} client_raw_response = models.OperationPaged(internal_paging, self._deserialize.dependencies, header_dict) return client_raw_response return deserialized
mit
-6,970,386,953,209,789,000
37.291667
144
0.616703
false
quantopian/zipline
tests/test_clock.py
1
5702
from datetime import time from unittest import TestCase import pandas as pd from trading_calendars import get_calendar from trading_calendars.utils.pandas_utils import days_at_time from zipline.gens.sim_engine import ( MinuteSimulationClock, SESSION_START, BEFORE_TRADING_START_BAR, BAR, SESSION_END ) class TestClock(TestCase): @classmethod def setUpClass(cls): cls.nyse_calendar = get_calendar("NYSE") # july 15 is friday, so there are 3 sessions in this range (15, 18, 19) cls.sessions = cls.nyse_calendar.sessions_in_range( pd.Timestamp("2016-07-15"), pd.Timestamp("2016-07-19") ) trading_o_and_c = cls.nyse_calendar.schedule.ix[cls.sessions] cls.opens = trading_o_and_c['market_open'] cls.closes = trading_o_and_c['market_close'] def test_bts_before_session(self): clock = MinuteSimulationClock( self.sessions, self.opens, self.closes, days_at_time(self.sessions, time(6, 17), "US/Eastern"), False ) all_events = list(clock) def _check_session_bts_first(session_label, events, bts_dt): minutes = self.nyse_calendar.minutes_for_session(session_label) self.assertEqual(393, len(events)) self.assertEqual(events[0], (session_label, SESSION_START)) self.assertEqual(events[1], (bts_dt, BEFORE_TRADING_START_BAR)) for i in range(2, 392): self.assertEqual(events[i], (minutes[i - 2], BAR)) self.assertEqual(events[392], (minutes[-1], SESSION_END)) _check_session_bts_first( self.sessions[0], all_events[0:393], pd.Timestamp("2016-07-15 6:17", tz='US/Eastern') ) _check_session_bts_first( self.sessions[1], all_events[393:786], pd.Timestamp("2016-07-18 6:17", tz='US/Eastern') ) _check_session_bts_first( self.sessions[2], all_events[786:], pd.Timestamp("2016-07-19 6:17", tz='US/Eastern') ) def test_bts_during_session(self): self.verify_bts_during_session( time(11, 45), [ pd.Timestamp("2016-07-15 11:45", tz='US/Eastern'), pd.Timestamp("2016-07-18 11:45", tz='US/Eastern'), pd.Timestamp("2016-07-19 11:45", tz='US/Eastern') ], 135 ) def test_bts_on_first_minute(self): self.verify_bts_during_session( time(9, 30), [ pd.Timestamp("2016-07-15 9:30", tz='US/Eastern'), pd.Timestamp("2016-07-18 9:30", tz='US/Eastern'), pd.Timestamp("2016-07-19 9:30", tz='US/Eastern') ], 1 ) def test_bts_on_last_minute(self): self.verify_bts_during_session( time(16, 00), [ pd.Timestamp("2016-07-15 16:00", tz='US/Eastern'), pd.Timestamp("2016-07-18 16:00", tz='US/Eastern'), pd.Timestamp("2016-07-19 16:00", tz='US/Eastern') ], 390 ) def verify_bts_during_session(self, bts_time, bts_session_times, bts_idx): def _check_session_bts_during(session_label, events, bts_dt): minutes = self.nyse_calendar.minutes_for_session(session_label) self.assertEqual(393, len(events)) self.assertEqual(events[0], (session_label, SESSION_START)) for i in range(1, bts_idx): self.assertEqual(events[i], (minutes[i - 1], BAR)) self.assertEqual( events[bts_idx], (bts_dt, BEFORE_TRADING_START_BAR) ) for i in range(bts_idx + 1, 391): self.assertEqual(events[i], (minutes[i - 2], BAR)) self.assertEqual(events[392], (minutes[-1], SESSION_END)) clock = MinuteSimulationClock( self.sessions, self.opens, self.closes, days_at_time(self.sessions, bts_time, "US/Eastern"), False ) all_events = list(clock) _check_session_bts_during( self.sessions[0], all_events[0:393], bts_session_times[0] ) _check_session_bts_during( self.sessions[1], all_events[393:786], bts_session_times[1] ) _check_session_bts_during( self.sessions[2], all_events[786:], bts_session_times[2] ) def test_bts_after_session(self): clock = MinuteSimulationClock( self.sessions, self.opens, self.closes, days_at_time(self.sessions, time(19, 5), "US/Eastern"), False ) all_events = list(clock) # since 19:05 Eastern is after the NYSE is closed, we don't emit # BEFORE_TRADING_START. therefore, each day has SESSION_START, # 390 BARs, and then SESSION_END def _check_session_bts_after(session_label, events): minutes = self.nyse_calendar.minutes_for_session(session_label) self.assertEqual(392, len(events)) self.assertEqual(events[0], (session_label, SESSION_START)) for i in range(1, 391): self.assertEqual(events[i], (minutes[i - 1], BAR)) self.assertEqual(events[-1], (minutes[389], SESSION_END)) for i in range(0, 2): _check_session_bts_after( self.sessions[i], all_events[(i * 392): ((i + 1) * 392)] )
apache-2.0
564,505,686,380,273,340
30.677778
79
0.539635
false
AutorestCI/azure-sdk-for-python
azure-mgmt-network/azure/mgmt/network/v2016_12_01/operations/routes_operations.py
1
17924
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- import uuid from msrest.pipeline import ClientRawResponse from msrestazure.azure_exceptions import CloudError from msrest.exceptions import DeserializationError from msrestazure.azure_operation import AzureOperationPoller from .. import models class RoutesOperations(object): """RoutesOperations operations. :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An objec model deserializer. :ivar api_version: Client API version. Constant value: "2016-12-01". """ models = models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self.api_version = "2016-12-01" self.config = config def _delete_initial( self, resource_group_name, route_table_name, route_name, custom_headers=None, raw=False, **operation_config): # Construct URL url = '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}' path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'routeTableName': self._serialize.url("route_table_name", route_table_name, 'str'), 'routeName': self._serialize.url("route_name", route_name, 'str'), 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request request = self._client.delete(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 202, 204]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response def delete( self, resource_group_name, route_table_name, route_name, custom_headers=None, raw=False, **operation_config): """Deletes the specified route from a route table. :param resource_group_name: The name of the resource group. :type resource_group_name: str :param route_table_name: The name of the route table. :type route_table_name: str :param route_name: The name of the route. :type route_name: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :return: An instance of AzureOperationPoller that returns None or ClientRawResponse if raw=true :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or ~msrest.pipeline.ClientRawResponse :raises: :class:`CloudError<msrestazure.azure_exceptions.CloudError>` """ raw_result = self._delete_initial( resource_group_name=resource_group_name, route_table_name=route_table_name, route_name=route_name, custom_headers=custom_headers, raw=True, **operation_config ) if raw: return raw_result # Construct and send request def long_running_send(): return raw_result.response def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) header_parameters = {} header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] return self._client.send( request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): if response.status_code not in [200, 202, 204]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout) def get( self, resource_group_name, route_table_name, route_name, custom_headers=None, raw=False, **operation_config): """Gets the specified route from a route table. :param resource_group_name: The name of the resource group. :type resource_group_name: str :param route_table_name: The name of the route table. :type route_table_name: str :param route_name: The name of the route. :type route_name: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: Route or ClientRawResponse if raw=true :rtype: ~azure.mgmt.network.v2016_12_01.models.Route or ~msrest.pipeline.ClientRawResponse :raises: :class:`CloudError<msrestazure.azure_exceptions.CloudError>` """ # Construct URL url = '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}' path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'routeTableName': self._serialize.url("route_table_name", route_table_name, 'str'), 'routeName': self._serialize.url("route_name", route_name, 'str'), 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp deserialized = None if response.status_code == 200: deserialized = self._deserialize('Route', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized def _create_or_update_initial( self, resource_group_name, route_table_name, route_name, route_parameters, custom_headers=None, raw=False, **operation_config): # Construct URL url = '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes/{routeName}' path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'routeTableName': self._serialize.url("route_table_name", route_table_name, 'str'), 'routeName': self._serialize.url("route_name", route_name, 'str'), 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body body_content = self._serialize.body(route_parameters, 'Route') # Construct and send request request = self._client.put(url, query_parameters) response = self._client.send( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200, 201]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp deserialized = None if response.status_code == 200: deserialized = self._deserialize('Route', response) if response.status_code == 201: deserialized = self._deserialize('Route', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized def create_or_update( self, resource_group_name, route_table_name, route_name, route_parameters, custom_headers=None, raw=False, **operation_config): """Creates or updates a route in the specified route table. :param resource_group_name: The name of the resource group. :type resource_group_name: str :param route_table_name: The name of the route table. :type route_table_name: str :param route_name: The name of the route. :type route_name: str :param route_parameters: Parameters supplied to the create or update route operation. :type route_parameters: ~azure.mgmt.network.v2016_12_01.models.Route :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :return: An instance of AzureOperationPoller that returns Route or ClientRawResponse if raw=true :rtype: ~msrestazure.azure_operation.AzureOperationPoller[~azure.mgmt.network.v2016_12_01.models.Route] or ~msrest.pipeline.ClientRawResponse :raises: :class:`CloudError<msrestazure.azure_exceptions.CloudError>` """ raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, route_table_name=route_table_name, route_name=route_name, route_parameters=route_parameters, custom_headers=custom_headers, raw=True, **operation_config ) if raw: return raw_result # Construct and send request def long_running_send(): return raw_result.response def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) header_parameters = {} header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] return self._client.send( request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): if response.status_code not in [200, 201]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp deserialized = self._deserialize('Route', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout) def list( self, resource_group_name, route_table_name, custom_headers=None, raw=False, **operation_config): """Gets all routes in a route table. :param resource_group_name: The name of the resource group. :type resource_group_name: str :param route_table_name: The name of the route table. :type route_table_name: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides<msrest:optionsforoperations>`. :return: An iterator like instance of Route :rtype: ~azure.mgmt.network.v2016_12_01.models.RoutePaged[~azure.mgmt.network.v2016_12_01.models.Route] :raises: :class:`CloudError<msrestazure.azure_exceptions.CloudError>` """ def internal_paging(next_link=None, raw=False): if not next_link: # Construct URL url = '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/routeTables/{routeTableName}/routes' path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'routeTableName': self._serialize.url("route_table_name", route_table_name, 'str'), 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str') } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') else: url = next_link query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request request = self._client.get(url, query_parameters) response = self._client.send( request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp return response # Deserialize response deserialized = models.RoutePaged(internal_paging, self._deserialize.dependencies) if raw: header_dict = {} client_raw_response = models.RoutePaged(internal_paging, self._deserialize.dependencies, header_dict) return client_raw_response return deserialized
mit
-8,097,506,177,059,138,000
43.81
158
0.632671
false
google-research/google-research
neural_guided_symbolic_regression/utils/generate_empirical_distribution_df_test.py
1
6620
# coding=utf-8 # Copyright 2021 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for generate_empirical_distribution_df.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from absl.testing import parameterized import numpy as np import pandas as pd import tensorflow.compat.v1 as tf from neural_guided_symbolic_regression.models import grammar_utils from neural_guided_symbolic_regression.utils import generate_empirical_distribution_df class GenerateEmpiricalDistributionDfHelperTest(parameterized.TestCase): @parameterized.parameters([ ([1, 4], 5), ([1, 4, 3], 5), ([1, 4, 3, 5, 2, 6], 3), ([1, 4, 3, 5, 2, 6, 8], 3) ]) def test_get_number_valid_next_step( self, prod_rules_sequence_indices, expected): grammar = grammar_utils.load_grammar( grammar_path='third_party/google_research/google_research/' 'neural_guided_symbolic_regression/grammar/' 'univariate_one_constant_grammar.txt') number_valid_next_step = ( generate_empirical_distribution_df.get_number_valid_next_step( prod_rules_sequence_indices, grammar)) self.assertEqual(number_valid_next_step, expected) class GenerateEmpiricalDistributionDfMainTest(parameterized.TestCase): def setUp(self): super(GenerateEmpiricalDistributionDfMainTest, self).setUp() # Production rule sequence of ( 1 ) is 1,6,7,6,9. # Production rule sequence of ( x ) is 1,6,7,6,8. self.expression_df = pd.DataFrame( {'expression_string': ['( 1 )', '( x )'], 'leading_at_0': [0, 1], 'leading_at_inf': [0, 1]}) self.grammar = grammar_utils.load_grammar( grammar_path='third_party/google_research/google_research/' 'neural_guided_symbolic_regression/grammar/' 'univariate_one_constant_grammar.txt') self.max_length = 11 def test_get_partial_sequence_df(self): partial_sequence_df = ( generate_empirical_distribution_df.get_partial_sequence_df( self.expression_df, self.grammar, self.max_length)) expected_partial_sequence_indices = ['1', '1_6', '1_6_7', '1_6_7_6', '1', '1_6', '1_6_7', '1_6_7_6'] self.assertListEqual( list(partial_sequence_df['partial_sequence_indices'].values), expected_partial_sequence_indices) @parameterized.parameters([ (None, 'partial_sequence_indices', ['1', 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0]), (None, 'partial_sequence_indices', ['1_6', 0, 0], [0, 0, 0, 0, 0, 0, 0, 1, 0, 0]), (None, 'partial_sequence_indices', ['1_6_7', 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0]), (None, 'partial_sequence_indices', ['1_6_7_6', 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 1]), (2, 'tail_partial_sequence_indices', ['1', 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0]), (2, 'tail_partial_sequence_indices', ['1_6', 0, 0], [0, 0, 0, 0, 0, 0, 0, 1, 0, 0]), (2, 'tail_partial_sequence_indices', ['6_7', 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0]), (2, 'tail_partial_sequence_indices', ['7_6', 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 1]), ]) def test_get_empirical_distribution_df(self, tail_length, level_name, multi_index_to_check, expected_probabilities): properties = ['leading_at_0', 'leading_at_inf'] num_production_rules = len(self.grammar.prod_rules) partial_sequence_df = ( generate_empirical_distribution_df.get_partial_sequence_df( self.expression_df, self.grammar, self.max_length)) empirical_distribution_df = ( generate_empirical_distribution_df.get_empirical_distribution_df( partial_sequence_df, properties, num_production_rules, tail_length)) levels = [level_name] + properties np.testing.assert_array_almost_equal( empirical_distribution_df.xs(multi_index_to_check, level=levels).values[0], expected_probabilities) def test_get_empirical_distribution_df_without_condition(self): num_production_rules = len(self.grammar.prod_rules) partial_sequence_df = ( generate_empirical_distribution_df.get_partial_sequence_df( self.expression_df, self.grammar, self.max_length)) empirical_distribution_df = ( generate_empirical_distribution_df.get_empirical_distribution_df( partial_sequence_df, [], num_production_rules, None)) expected = pd.DataFrame( np.array([[0, 0, 0, 0, 0, 0, 1, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 1, 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0.5, 0.5]]), columns=range(10)) expected['partial_sequence_indices'] = ['1', '1_6', '1_6_7', '1_6_7_6'] expected.set_index('partial_sequence_indices', inplace=True) pd.testing.assert_frame_equal(empirical_distribution_df, expected, check_dtype=False, check_index_type=False, check_column_type=False, check_names=False) @parameterized.parameters([ ('1_6_7_6', 1, '6'), ('1_6_7_6', 2, '7_6'), ('1_6', 3, '1_6'), ]) def test_extract_tail_partial_sequence(self, partial_sequence_string, tail_length, expected): tail_partial_sequence_string = ( generate_empirical_distribution_df.extract_tail_partial_sequence( partial_sequence_string, tail_length)) self.assertEqual(tail_partial_sequence_string, expected) if __name__ == '__main__': tf.test.main()
apache-2.0
-8,822,368,325,674,619,000
37.71345
86
0.575378
false
p-montero/py-ans
class8/ex7.py
1
1695
#!/usr/bin/env python ''' Use processes and Netmiko to connect to each of the devices in the database. Execute 'show version' on each device. Record the amount of time required to do this. DISCLAIMER NOTE: Solution is limited to the exercise's scope ''' from net_system.models import NetworkDevice import django from multiprocessing import Process from termcolor import colored from datetime import datetime from netmiko import ConnectHandler def sh_ver(a_device): # Execute cmd with NETMIKO creds = a_device.credentials rem_conn_ssh = ConnectHandler(device_type=a_device.device_type, ip=a_device.ip_address, username=creds.username, password=creds.password, port=a_device.port, secret='') # Output cmd output = rem_conn_ssh.send_command_expect("show version") print "\n <<--------------------------->> \n "+ colored(output, 'green') + "\n" def main(): # Main function to connect to the devices using NETMIKO and execute a cmd. Multi-processing support. django.setup() # Record start time process = [] start_time = datetime.now() pylab_devices = NetworkDevice.objects.all() for a_device in pylab_devices: # Create a PROCESS for each device connection/cmd node_process = Process(target=sh_ver, args=(a_device,)) # Start the THREAD node_process.start() process.append(node_process) for any_process in process: print "Notice: " + colored(any_process, 'red') any_process.join() # Function sh_ver runtime calculation runtime = datetime.now() - start_time print "This operation required " + colored(runtime, 'blue') if __name__ == "__main__": main()
apache-2.0
5,557,594,805,842,516,000
35.06383
116
0.676106
false
disco-stu/pcs
pcs/test/test_utils.py
1
58548
import os import sys import shutil import unittest import xml.dom.minidom parentdir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.insert(0, parentdir) from pcs_test_functions import pcs, ac import utils empty_cib = "empty.xml" temp_cib = "temp.xml" class UtilsTest(unittest.TestCase): def get_cib_empty(self): return xml.dom.minidom.parse("empty.xml") def get_cib_resources(self): cib_dom = self.get_cib_empty() new_resources = xml.dom.minidom.parseString(""" <resources> <primitive id="myResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> <clone id="myClone"> <primitive id="myClonedResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> </clone> <master id="myMaster"> <primitive id="myMasteredResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> </master> <group id="myGroup"> <primitive id="myGroupedResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> </group> <clone id="myGroupClone"> <group id="myClonedGroup"> <primitive id="myClonedGroupedResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> </group> </clone> <master id="myGroupMaster"> <group id="myMasteredGroup"> <primitive id="myMasteredGroupedResource" class="ocf" provider="heartbeat" type="Dummy"> </primitive> </group> </master> </resources> """).documentElement resources = cib_dom.getElementsByTagName("resources")[0] resources.parentNode.replaceChild(new_resources, resources) return cib_dom def testDomGetResources(self): def test_dom_get(method, dom, ok_ids, bad_ids): for element_id in ok_ids: self.assert_element_id(method(dom, element_id), element_id) for element_id in bad_ids: self.assertFalse(method(dom, element_id)) cib_dom = self.get_cib_empty() self.assertFalse(utils.dom_get_resource(cib_dom, "myResource")) self.assertFalse( utils.dom_get_resource_clone(cib_dom, "myClonedResource") ) self.assertFalse( utils.dom_get_resource_masterslave(cib_dom, "myMasteredResource") ) self.assertFalse(utils.dom_get_group(cib_dom, "myGroup")) self.assertFalse(utils.dom_get_group_clone(cib_dom, "myClonedGroup")) self.assertFalse( utils.dom_get_group_masterslave(cib_dom, "myMasteredGroup") ) self.assertFalse(utils.dom_get_clone(cib_dom, "myClone")) self.assertFalse(utils.dom_get_master(cib_dom, "myMaster")) self.assertFalse(utils.dom_get_clone_ms_resource(cib_dom, "myClone")) self.assertFalse(utils.dom_get_clone_ms_resource(cib_dom, "myMaster")) self.assertFalse( utils.dom_get_resource_clone_ms_parent(cib_dom, "myClonedResource") ) self.assertFalse( utils.dom_get_resource_clone_ms_parent(cib_dom, "myMasteredResource") ) cib_dom = self.get_cib_resources() all_ids = set([ "none", "myResource", "myClone", "myClonedResource", "myMaster", "myMasteredResource", "myGroup", "myGroupedResource", "myGroupClone", "myClonedGroup", "myClonedGroupedResource", "myGroupMaster", "myMasteredGroup", "myMasteredGroupedResource", ]) resource_ids = set([ "myResource", "myClonedResource", "myGroupedResource", "myMasteredResource", "myClonedGroupedResource", "myMasteredGroupedResource" ]) test_dom_get( utils.dom_get_resource, cib_dom, resource_ids, all_ids - resource_ids ) cloned_ids = set(["myClonedResource", "myClonedGroupedResource"]) test_dom_get( utils.dom_get_resource_clone, cib_dom, cloned_ids, all_ids - cloned_ids ) mastered_ids = set(["myMasteredResource", "myMasteredGroupedResource"]) test_dom_get( utils.dom_get_resource_masterslave, cib_dom, mastered_ids, all_ids - mastered_ids ) group_ids = set(["myGroup", "myClonedGroup", "myMasteredGroup"]) test_dom_get( utils.dom_get_group, cib_dom, group_ids, all_ids - group_ids ) cloned_group_ids = set(["myClonedGroup"]) test_dom_get( utils.dom_get_group_clone, cib_dom, cloned_group_ids, all_ids - cloned_group_ids ) clone_ids = set(["myClone", "myGroupClone"]) test_dom_get( utils.dom_get_clone, cib_dom, clone_ids, all_ids - clone_ids ) mastered_group_ids = set(["myMasteredGroup"]) test_dom_get( utils.dom_get_group_masterslave, cib_dom, mastered_group_ids, all_ids - mastered_group_ids ) master_ids = set(["myMaster", "myGroupMaster"]) test_dom_get( utils.dom_get_master, cib_dom, master_ids, all_ids - master_ids ) self.assert_element_id( utils.dom_get_clone_ms_resource(cib_dom, "myClone"), "myClonedResource" ) self.assert_element_id( utils.dom_get_clone_ms_resource(cib_dom, "myGroupClone"), "myClonedGroup" ) self.assert_element_id( utils.dom_get_clone_ms_resource(cib_dom, "myMaster"), "myMasteredResource" ) self.assert_element_id( utils.dom_get_clone_ms_resource(cib_dom, "myGroupMaster"), "myMasteredGroup" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent(cib_dom, "myClonedResource"), "myClone" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent(cib_dom, "myClonedGroup"), "myGroupClone" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent( cib_dom, "myClonedGroupedResource" ), "myGroupClone" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent( cib_dom, "myMasteredResource" ), "myMaster" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent( cib_dom, "myMasteredGroup" ), "myGroupMaster" ) self.assert_element_id( utils.dom_get_resource_clone_ms_parent( cib_dom, "myMasteredGroupedResource" ), "myGroupMaster" ) self.assertEquals( None, utils.dom_get_resource_clone_ms_parent(cib_dom, "myResource") ) self.assertEquals( None, utils.dom_get_resource_clone_ms_parent(cib_dom, "myGroup") ) self.assertEquals( None, utils.dom_get_resource_clone_ms_parent(cib_dom, "myGroupedResource") ) def testDomGetResourceRemoteNodeName(self): dom = xml.dom.minidom.parse("empty.xml") new_resources = xml.dom.minidom.parseString(""" <resources> <primitive id="dummy1" class="ocf" provider="heartbeat" type="Dummy"> </primitive> <primitive class="ocf" id="vm-guest1" provider="heartbeat" type="VirtualDomain"> <instance_attributes id="vm-guest1-instance_attributes"> <nvpair id="vm-guest1-instance_attributes-hypervisor" name="hypervisor" value="qemu:///system"/> <nvpair id="vm-guest1-instance_attributes-config" name="config" value="/root/guest1.xml"/> </instance_attributes> <meta_attributes id="vm-guest1-meta_attributes"> <nvpair id="vm-guest1-meta_attributes-remote-node" name="remote-node" value="guest1"/> </meta_attributes> </primitive> <primitive id="dummy2" class="ocf" provider="heartbeat" type="Dummy"> <instance_attributes id="vm-guest1-meta_attributes"> <nvpair id="dummy2-remote-node" name="remote-node" value="guest2"/> </instance_attributes> </primitive> </resources> """).documentElement resources = dom.getElementsByTagName("resources")[0] resources.parentNode.replaceChild(new_resources, resources) self.assertEquals( None, utils.dom_get_resource_remote_node_name( utils.dom_get_resource(dom, "dummy1") ) ) self.assertEquals( None, utils.dom_get_resource_remote_node_name( utils.dom_get_resource(dom, "dummy2") ) ) self.assertEquals( "guest1", utils.dom_get_resource_remote_node_name( utils.dom_get_resource(dom, "vm-guest1") ) ) def test_dom_get_meta_attr_value(self): dom = xml.dom.minidom.parse("empty.xml") new_resources = xml.dom.minidom.parseString(""" <resources> <primitive id="dummy1" class="ocf" provider="heartbeat" type="Dummy"> </primitive> <primitive class="ocf" id="vm-guest1" provider="heartbeat" type="VirtualDomain"> <instance_attributes id="vm-guest1-instance_attributes"> <nvpair id="vm-guest1-instance_attributes-hypervisor" name="hypervisor" value="qemu:///system"/> <nvpair id="vm-guest1-instance_attributes-config" name="config" value="/root/guest1.xml"/> </instance_attributes> <meta_attributes id="vm-guest1-meta_attributes"> <nvpair id="vm-guest1-meta_attributes-remote-node" name="remote-node" value="guest1"/> </meta_attributes> </primitive> <primitive id="dummy2" class="ocf" provider="heartbeat" type="Dummy"> <instance_attributes id="vm-guest1-meta_attributes"> <nvpair id="dummy2-remote-node" name="remote-node" value="guest2"/> </instance_attributes> </primitive> </resources> """).documentElement resources = dom.getElementsByTagName("resources")[0] resources.parentNode.replaceChild(new_resources, resources) self.assertEquals( None, utils.dom_get_meta_attr_value( utils.dom_get_resource(dom, "dummy1"), "foo" ) ) self.assertEquals( None, utils.dom_get_meta_attr_value( utils.dom_get_resource(dom, "dummy2"), "remote-node" ) ) self.assertEquals( "guest1", utils.dom_get_meta_attr_value( utils.dom_get_resource(dom, "vm-guest1"), "remote-node" ) ) self.assertEquals( None, utils.dom_get_meta_attr_value( utils.dom_get_resource(dom, "vm-guest1"), "foo" ) ) def testGetElementWithId(self): dom = xml.dom.minidom.parseString(""" <aa> <bb id="bb1"/> <bb/> <bb id="bb2"> <cc id="cc1"/> </bb> <bb id="bb3"> <cc id="cc2"/> </bb> </aa> """).documentElement self.assert_element_id( utils.dom_get_element_with_id(dom, "bb", "bb1"), "bb1" ) self.assert_element_id( utils.dom_get_element_with_id(dom, "bb", "bb2"), "bb2" ) self.assert_element_id( utils.dom_get_element_with_id(dom, "cc", "cc1"), "cc1" ) self.assert_element_id( utils.dom_get_element_with_id( utils.dom_get_element_with_id(dom, "bb", "bb2"), "cc", "cc1" ), "cc1" ) self.assertEquals(None, utils.dom_get_element_with_id(dom, "dd", "bb1")) self.assertEquals(None, utils.dom_get_element_with_id(dom, "bb", "bb4")) self.assertEquals(None, utils.dom_get_element_with_id(dom, "bb", "cc1")) self.assertEquals( None, utils.dom_get_element_with_id( utils.dom_get_element_with_id(dom, "bb", "bb2"), "cc", "cc2" ) ) def test_dom_get_parent_by_tag_name(self): dom = xml.dom.minidom.parseString(""" <aa id="aa1"> <bb id="bb1"/> <bb id="bb2"> <cc id="cc1"/> </bb> <bb id="bb3"> <cc id="cc2"/> </bb> <dd id="dd1" /> </aa> """).documentElement bb1 = utils.dom_get_element_with_id(dom, "bb", "bb1") cc1 = utils.dom_get_element_with_id(dom, "cc", "cc1") self.assert_element_id( utils.dom_get_parent_by_tag_name(bb1, "aa"), "aa1" ) self.assert_element_id( utils.dom_get_parent_by_tag_name(cc1, "aa"), "aa1" ) self.assert_element_id( utils.dom_get_parent_by_tag_name(cc1, "bb"), "bb2" ) self.assertEquals(None, utils.dom_get_parent_by_tag_name(bb1, "cc")) self.assertEquals(None, utils.dom_get_parent_by_tag_name(cc1, "dd")) self.assertEquals(None, utils.dom_get_parent_by_tag_name(cc1, "ee")) def testValidateConstraintResource(self): dom = self.get_cib_resources() self.assertEquals( (True, "", "myClone"), utils.validate_constraint_resource(dom, "myClone") ) self.assertEquals( (True, "", "myGroupClone"), utils.validate_constraint_resource(dom, "myGroupClone") ) self.assertEquals( (True, "", "myMaster"), utils.validate_constraint_resource(dom, "myMaster") ) self.assertEquals( (True, "", "myGroupMaster"), utils.validate_constraint_resource(dom, "myGroupMaster") ) self.assertEquals( (True, "", "myResource"), utils.validate_constraint_resource(dom, "myResource") ) self.assertEquals( (True, "", "myGroup"), utils.validate_constraint_resource(dom, "myGroup") ) self.assertEquals( (True, "", "myGroupedResource"), utils.validate_constraint_resource(dom, "myGroupedResource") ) self.assertEquals( (False, "Resource 'myNonexistent' does not exist", None), utils.validate_constraint_resource(dom, "myNonexistent") ) message = ( "%s is a clone resource, you should use the clone id: " "%s when adding constraints. Use --force to override." ) self.assertEquals( ( False, message % ("myClonedResource", "myClone"), "myClone" ), utils.validate_constraint_resource(dom, "myClonedResource") ) self.assertEquals( ( False, message % ("myClonedGroup", "myGroupClone"), "myGroupClone" ), utils.validate_constraint_resource(dom, "myClonedGroup") ) self.assertEquals( ( False, message % ("myClonedGroupedResource", "myGroupClone"), "myGroupClone" ), utils.validate_constraint_resource(dom, "myClonedGroupedResource") ) message = ( "%s is a master/slave resource, you should use the master id: " "%s when adding constraints. Use --force to override." ) self.assertEquals( ( False, message % ("myMasteredResource", "myMaster"), "myMaster" ), utils.validate_constraint_resource(dom, "myMasteredResource") ) self.assertEquals( ( False, message % ("myMasteredGroup", "myGroupMaster"), "myGroupMaster" ), utils.validate_constraint_resource(dom, "myMasteredGroup") ) self.assertEquals( ( False, message % ("myMasteredGroupedResource", "myGroupMaster"), "myGroupMaster" ), utils.validate_constraint_resource(dom, "myMasteredGroupedResource") ) utils.pcs_options["--force"] = True self.assertEquals( (True, "", "myClone"), utils.validate_constraint_resource(dom, "myClonedResource") ) self.assertEquals( (True, "", "myGroupClone"), utils.validate_constraint_resource(dom, "myClonedGroup") ) self.assertEquals( (True, "", "myGroupClone"), utils.validate_constraint_resource(dom, "myClonedGroupedResource") ) self.assertEquals( (True, "", "myMaster"), utils.validate_constraint_resource(dom, "myMasteredResource") ) self.assertEquals( (True, "", "myGroupMaster"), utils.validate_constraint_resource(dom, "myMasteredGroup") ) self.assertEquals( (True, "", "myGroupMaster"), utils.validate_constraint_resource(dom, "myMasteredGroupedResource") ) def testValidateXmlId(self): self.assertEquals((True, ""), utils.validate_xml_id("dummy")) self.assertEquals((True, ""), utils.validate_xml_id("DUMMY")) self.assertEquals((True, ""), utils.validate_xml_id("dUmMy")) self.assertEquals((True, ""), utils.validate_xml_id("dummy0")) self.assertEquals((True, ""), utils.validate_xml_id("dum0my")) self.assertEquals((True, ""), utils.validate_xml_id("dummy-")) self.assertEquals((True, ""), utils.validate_xml_id("dum-my")) self.assertEquals((True, ""), utils.validate_xml_id("dummy.")) self.assertEquals((True, ""), utils.validate_xml_id("dum.my")) self.assertEquals((True, ""), utils.validate_xml_id("_dummy")) self.assertEquals((True, ""), utils.validate_xml_id("dummy_")) self.assertEquals((True, ""), utils.validate_xml_id("dum_my")) self.assertEquals( (False, "test id cannot be empty"), utils.validate_xml_id("", "test id") ) msg = "invalid test id '%s', '%s' is not a valid first character for a test id" self.assertEquals( (False, msg % ("0", "0")), utils.validate_xml_id("0", "test id") ) self.assertEquals( (False, msg % ("-", "-")), utils.validate_xml_id("-", "test id") ) self.assertEquals( (False, msg % (".", ".")), utils.validate_xml_id(".", "test id") ) self.assertEquals( (False, msg % (":", ":")), utils.validate_xml_id(":", "test id") ) self.assertEquals( (False, msg % ("0dummy", "0")), utils.validate_xml_id("0dummy", "test id") ) self.assertEquals( (False, msg % ("-dummy", "-")), utils.validate_xml_id("-dummy", "test id") ) self.assertEquals( (False, msg % (".dummy", ".")), utils.validate_xml_id(".dummy", "test id") ) self.assertEquals( (False, msg % (":dummy", ":")), utils.validate_xml_id(":dummy", "test id") ) msg = "invalid test id '%s', '%s' is not a valid character for a test id" self.assertEquals( (False, msg % ("dum:my", ":")), utils.validate_xml_id("dum:my", "test id") ) self.assertEquals( (False, msg % ("dummy:", ":")), utils.validate_xml_id("dummy:", "test id") ) self.assertEquals( (False, msg % ("dum?my", "?")), utils.validate_xml_id("dum?my", "test id") ) self.assertEquals( (False, msg % ("dummy?", "?")), utils.validate_xml_id("dummy?", "test id") ) def testIsIso8601Date(self): self.assertTrue(utils.is_iso8601_date("2014-07-03")) self.assertTrue(utils.is_iso8601_date("2014-07-03T11:35:14")) self.assertTrue(utils.is_iso8601_date("20140703")) self.assertTrue(utils.is_iso8601_date("2014-W27-4")) self.assertTrue(utils.is_iso8601_date("2014-184")) self.assertFalse(utils.is_iso8601_date("")) self.assertFalse(utils.is_iso8601_date("foo")) self.assertFalse(utils.is_iso8601_date("2014-07-32")) self.assertFalse(utils.is_iso8601_date("2014-13-03")) self.assertFalse(utils.is_iso8601_date("2014-W27-8")) self.assertFalse(utils.is_iso8601_date("2014-367")) def test_is_score(self): self.assertTrue(utils.is_score("INFINITY")) self.assertTrue(utils.is_score("+INFINITY")) self.assertTrue(utils.is_score("-INFINITY")) self.assertTrue(utils.is_score("0")) self.assertTrue(utils.is_score("+0")) self.assertTrue(utils.is_score("-0")) self.assertTrue(utils.is_score("123")) self.assertTrue(utils.is_score("-123")) self.assertTrue(utils.is_score("+123")) self.assertFalse(utils.is_score("")) self.assertFalse(utils.is_score("abc")) self.assertFalse(utils.is_score("+abc")) self.assertFalse(utils.is_score("-abc")) self.assertFalse(utils.is_score("10a")) self.assertFalse(utils.is_score("+10a")) self.assertFalse(utils.is_score("-10a")) self.assertFalse(utils.is_score("a10")) self.assertFalse(utils.is_score("+a10")) self.assertFalse(utils.is_score("a-10")) self.assertFalse(utils.is_score("infinity")) self.assertFalse(utils.is_score("+infinity")) self.assertFalse(utils.is_score("-infinity")) self.assertFalse(utils.is_score("+InFiNiTy")) self.assertFalse(utils.is_score("INFINITY10")) self.assertFalse(utils.is_score("INFINITY+10")) self.assertFalse(utils.is_score("-INFINITY10")) self.assertFalse(utils.is_score("+INFINITY+10")) self.assertFalse(utils.is_score("10INFINITY")) self.assertFalse(utils.is_score("+10+INFINITY")) def test_get_timeout_seconds(self): self.assertEquals(utils.get_timeout_seconds("10"), 10) self.assertEquals(utils.get_timeout_seconds("10s"), 10) self.assertEquals(utils.get_timeout_seconds("10min"), 600) self.assertEquals(utils.get_timeout_seconds("1a1s"), None) self.assertEquals(utils.get_timeout_seconds("10m"), None) self.assertEquals(utils.get_timeout_seconds("10mim"), None) self.assertEquals(utils.get_timeout_seconds("aaa"), None) self.assertEquals(utils.get_timeout_seconds(""), None) self.assertEquals(utils.get_timeout_seconds("1a1s", True), "1a1s") self.assertEquals(utils.get_timeout_seconds("10m", True), "10m") self.assertEquals(utils.get_timeout_seconds("10mim", True), "10mim") self.assertEquals(utils.get_timeout_seconds("aaa", True), "aaa") self.assertEquals(utils.get_timeout_seconds("", True), "") def test_get_default_op_timeout(self): shutil.copy(empty_cib, temp_cib) utils.usefile = True utils.filename = temp_cib self.assertEquals(utils.get_default_op_timeout(), 20) output, retVal = pcs(temp_cib, "property set default-action-timeout=25") self.assertEquals(retVal, 0) self.assertEquals(utils.get_default_op_timeout(), 25) output, retVal = pcs(temp_cib, "property unset default-action-timeout") self.assertEquals(retVal, 0) self.assertEquals(utils.get_default_op_timeout(), 20) utils.usefile = False utils.filename = "" def test_get_resource_op_timeout(self): shutil.copy(empty_cib, temp_cib) utils.usefile = True utils.filename = temp_cib output, retVal = pcs(temp_cib, "property set default-action-timeout=25") ac(output, "") self.assertEquals(retVal, 0) output, retVal = pcs( temp_cib, "resource create dummy Dummy op start timeout=33s --no-default-ops" ) ac(output, "") self.assertEquals(retVal, 0) dom = xml.dom.minidom.parse(temp_cib) self.assertEquals( utils.get_resource_op_timeout(dom, "dummy", "start"), 33 ) self.assertEquals( utils.get_resource_op_timeout(dom, "dummy", "stop"), 20 ) self.assertEquals( utils.get_resource_op_timeout(dom, "dummy0", "start"), 25 ) utils.usefile = False utils.filename = "" def get_cib_status_lrm(self): cib_dom = self.get_cib_empty() new_status = xml.dom.minidom.parseString(""" <status> <node_state id="1" uname="rh70-node1"> <lrm id="1"> <lrm_resources> <lrm_resource id="dummy" type="Dummy" class="ocf" provider="heartbeat"> <lrm_rsc_op id="dummy_monitor_30000" operation="monitor" call-id="34" rc-code="1" on_node="Xrh70-node1X" exit-reason="test" /> <lrm_rsc_op id="dummy_stop_0" operation="stop" call-id="32" rc-code="0" /> <lrm_rsc_op id="dummy_start_0" operation="start" call-id="33" rc-code="0" /> </lrm_resource> </lrm_resources> </lrm> </node_state> <node_state id="2" uname="rh70-node2"> <lrm id="2"> <lrm_resources> <lrm_resource id="dummy" type="Dummy" class="ocf" provider="heartbeat"> <lrm_rsc_op id="dummy_monitor_0" operation="monitor" call-id="5" rc-code="1" /> </lrm_resource> <lrm_resource id="dummy1" type="Dummy" class="ocf" provider="heartbeat"> <lrm_rsc_op id="dummy1_monitor_0" operation="monitor" call-id="3" rc-code="0" /> </lrm_resource> </lrm_resources> </lrm> </node_state> </status> """).documentElement status = cib_dom.getElementsByTagName("status")[0] status.parentNode.replaceChild(new_status, status) return cib_dom def test_get_lrm_rsc_op(self): dom = self.get_cib_status_lrm() op_list = utils.get_lrm_rsc_op(dom, "dummy") op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals( op_id_list, ["dummy_monitor_0", "dummy_stop_0", "dummy_start_0", "dummy_monitor_30000",] ) op_list = utils.get_lrm_rsc_op(dom, "dummy", ["monitor"]) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals( op_id_list, ["dummy_monitor_0", "dummy_monitor_30000",] ) op_list = utils.get_lrm_rsc_op(dom, "dummy", ["stop", "start"]) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals( op_id_list, ["dummy_stop_0", "dummy_start_0",] ) op_list = utils.get_lrm_rsc_op(dom, "dummy", last_call_id=30) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals( op_id_list, ["dummy_stop_0", "dummy_start_0", "dummy_monitor_30000",] ) op_list = utils.get_lrm_rsc_op(dom, "dummy", ["monitor"], 30) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals( op_id_list, ["dummy_monitor_30000",] ) op_list = utils.get_lrm_rsc_op(dom, "dummy", last_call_id=340) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) op_list = utils.get_lrm_rsc_op(dom, "dummy", last_call_id=34) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) op_list = utils.get_lrm_rsc_op(dom, "dummy0", ["monitor"], 30) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) op_list = utils.get_lrm_rsc_op(dom, "dummy0", ["monitor"]) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) op_list = utils.get_lrm_rsc_op(dom, "dummy0", last_call_id=30) op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) op_list = utils.get_lrm_rsc_op(dom, "dummy0") op_id_list = [op.getAttribute("id") for op in op_list] self.assertEquals(op_id_list, []) def test_get_lrm_rsc_op_failures(self): dom = self.get_cib_status_lrm() failures = utils.get_lrm_rsc_op_failures( utils.get_lrm_rsc_op(dom, "dummy") ) self.assertEquals( failures, ["rh70-node2: failed", "Xrh70-node1X: test"] ) failures = utils.get_lrm_rsc_op_failures( utils.get_lrm_rsc_op(dom, "dummy", ["start"]) ) self.assertEquals(failures, []) failures = utils.get_lrm_rsc_op_failures( utils.get_lrm_rsc_op(dom, "dummy0") ) self.assertEquals(failures, []) def test_resource_running_on(self): status = xml.dom.minidom.parseString(""" <crm_mon> <summary /> <nodes /> <resources> <resource id="myResource" role="Started"> <node name="rh70-node1" /> </resource> <clone id="myClone"> <resource id="myClonedResource" role="Started"> <node name="rh70-node1" /> </resource> <resource id="myClonedResource" role="Started"> <node name="rh70-node2" /> </resource> <resource id="myClonedResource" role="Started"> <node name="rh70-node3" /> </resource> </clone> <clone id="myMaster"> <resource id="myMasteredResource:1" role="Slave"> <node name="rh70-node2" /> </resource> <resource id="myMasteredResource" role="Slave"> <node name="rh70-node3" /> </resource> <resource id="myMasteredResource" role="Master"> <node name="rh70-node1" /> </resource> </clone> <group id="myGroup"> <resource id="myGroupedResource" role="Started"> <node name="rh70-node2" /> </resource> </group> <clone id="myGroupClone"> <group id="myClonedGroup:0"> <resource id="myClonedGroupedResource" role="Started"> <node name="rh70-node1" /> </resource> </group> <group id="myClonedGroup:1"> <resource id="myClonedGroupedResource" role="Started"> <node name="rh70-node2" /> </resource> </group> <group id="myClonedGroup:2"> <resource id="myClonedGroupedResource" role="Started"> <node name="rh70-node3" /> </resource> </group> <group id="myClonedGroup:3"> <resource id="myClonedGroupedResource" role="Started"> <node name="rh70-node3" /> </resource> </group> </clone> <clone id="myGroupMaster"> <group id="myMasteredGroup:0"> <resource id="myMasteredGroupedResource" role="Slave"> <node name="rh70-node1" /> </resource> </group> <group id="myMasteredGroup:1"> <resource id="myMasteredGroupedResource" role="Master"> <node name="rh70-node2" /> </resource> </group> <group id="myMasteredGroup:2"> <resource id="myMasteredGroupedResource" role="Slave"> <node name="rh70-node3" /> </resource> </group> </clone> <resource id="myStoppedResource" role="Stopped"> </resource> </resources> </crm_mon> """).documentElement self.assertEquals( utils.resource_running_on("myResource", status), { 'message': "Resource 'myResource' is running on node rh70-node1.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1"], } ) self.assertEquals( utils.resource_running_on("myClonedResource", status), { 'message': "Resource 'myClonedResource' is running on nodes " "rh70-node1, rh70-node2, rh70-node3.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1", "rh70-node2", "rh70-node3"], } ) self.assertEquals( utils.resource_running_on("myClone", status), { 'message': "Resource 'myClone' is running on nodes " "rh70-node1, rh70-node2, rh70-node3.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1", "rh70-node2", "rh70-node3"], } ) self.assertEquals( utils.resource_running_on("myMasteredResource", status), { 'message': "Resource 'myMasteredResource' is master on node " "rh70-node1; slave on nodes rh70-node2, rh70-node3.", 'nodes_master': ["rh70-node1"], 'nodes_slave': ["rh70-node2", "rh70-node3"], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("myMaster", status), { 'message': "Resource 'myMaster' is master on node " "rh70-node1; slave on nodes rh70-node2, rh70-node3.", 'nodes_master': ["rh70-node1"], 'nodes_slave': ["rh70-node2", "rh70-node3"], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("myGroupedResource", status), { 'message': "Resource 'myGroupedResource' is running on node " "rh70-node2.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node2"], } ) self.assertEquals( utils.resource_running_on("myGroup", status), { 'message': "Resource 'myGroup' is running on node " "rh70-node2.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node2"], } ) self.assertEquals( utils.resource_running_on("myClonedGroupedResource", status), { 'message': "Resource 'myClonedGroupedResource' is running on nodes " "rh70-node1, rh70-node2, rh70-node3, rh70-node3.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1", "rh70-node2", "rh70-node3", "rh70-node3"], } ) self.assertEquals( utils.resource_running_on("myClonedGroup", status), { 'message': "Resource 'myClonedGroup' is running on nodes " "rh70-node1, rh70-node2, rh70-node3, rh70-node3.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1", "rh70-node2", "rh70-node3", "rh70-node3"], } ) self.assertEquals( utils.resource_running_on("myGroupClone", status), { 'message': "Resource 'myGroupClone' is running on nodes " "rh70-node1, rh70-node2, rh70-node3, rh70-node3.", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': ["rh70-node1", "rh70-node2", "rh70-node3", "rh70-node3"], } ) self.assertEquals( utils.resource_running_on("myMasteredGroupedResource", status), { 'message': "Resource 'myMasteredGroupedResource' is master on node " "rh70-node2; slave on nodes rh70-node1, rh70-node3.", 'nodes_master': ["rh70-node2"], 'nodes_slave': ["rh70-node1", "rh70-node3"], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("myMasteredGroup", status), { 'message': "Resource 'myMasteredGroup' is master on node " "rh70-node2; slave on nodes rh70-node1, rh70-node3.", 'nodes_master': ["rh70-node2"], 'nodes_slave': ["rh70-node1", "rh70-node3"], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("myGroupMaster", status), { 'message': "Resource 'myGroupMaster' is master on node " "rh70-node2; slave on nodes rh70-node1, rh70-node3.", 'nodes_master': ["rh70-node2"], 'nodes_slave': ["rh70-node1", "rh70-node3"], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("notMyResource", status), { 'message': "Resource 'notMyResource' is not running on any node", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': [], } ) self.assertEquals( utils.resource_running_on("myStoppedResource", status), { 'message': "Resource 'myStoppedResource' is not running on any node", 'nodes_master': [], 'nodes_slave': [], 'nodes_started': [], } ) def test_count_expected_resource_instances(self): dom = xml.dom.minidom.parse("empty.xml") new_resources = xml.dom.minidom.parseString(""" <resources> <primitive id="prim1"> </primitive> <group id="group1"> <primitive id="prim2"> </primitive> </group> <clone id="clone1"> <primitive id="prim3"> </primitive> </clone> <clone id="clone2"> <primitive id="prim4"> </primitive> <meta_attributes> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="3"/> </meta_attributes> </clone> <clone id="clone3"> <primitive id="prim5"> </primitive> <meta_attributes> <nvpair name="clone-max" value="2"/> <nvpair name="clone-node-max" value="3"/> </meta_attributes> </clone> <clone id="clone4"> <primitive id="prim6"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> </meta_attributes> </clone> <clone id="clone5"> <primitive id="prim7"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="2"/> </meta_attributes> </clone> <clone id="clone6"> <primitive id="prim8"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="4"/> </meta_attributes> </clone> <master id="master1"> <primitive id="prim9"> </primitive> </master> <master id="master2"> <primitive id="prim10"> </primitive> <meta_attributes> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="3"/> <nvpair name="master-max" value="5"/> <nvpair name="master-node-max" value="4"/> </meta_attributes> </master> <master id="master3"> <primitive id="prim11"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="3"/> </meta_attributes> </master> <master id="master4"> <primitive id="prim12"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="3"/> <nvpair name="master-max" value="3"/> <nvpair name="master-node-max" value="2"/> </meta_attributes> </master> <master id="master5"> <primitive id="prim13"> </primitive> <meta_attributes> <nvpair name="globally-unique" value="true"/> <nvpair name="clone-max" value="9"/> <nvpair name="clone-node-max" value="3"/> <nvpair name="master-max" value="12"/> <nvpair name="master-node-max" value="4"/> </meta_attributes> </master> </resources> """).documentElement resources = dom.getElementsByTagName("resources")[0] resources.parentNode.replaceChild(new_resources, resources) self.assertEquals( 1, utils.count_expected_resource_instances( utils.dom_get_resource(dom, "prim1"), 3 ) ) self.assertEquals( 1, utils.count_expected_resource_instances( utils.dom_get_group(dom, "group1"), 3 ) ) self.assertEquals( 3, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone1"), 3 ) ) self.assertEquals( 3, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone2"), 3 ) ) self.assertEquals( 2, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone3"), 3 ) ) self.assertEquals( 3, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone4"), 3 ) ) self.assertEquals( 6, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone5"), 3 ) ) self.assertEquals( 9, utils.count_expected_resource_instances( utils.dom_get_clone(dom, "clone6"), 3 ) ) self.assertEquals( 1, utils.count_expected_resource_instances( utils.dom_get_master(dom, "master1"), 3 ) ) self.assertEquals( 3, utils.count_expected_resource_instances( utils.dom_get_master(dom, "master2"), 3 ) ) self.assertEquals( 1, utils.count_expected_resource_instances( utils.dom_get_master(dom, "master3"), 3 ) ) self.assertEquals( 3, utils.count_expected_resource_instances( utils.dom_get_master(dom, "master4"), 3 ) ) self.assertEquals( 9, utils.count_expected_resource_instances( utils.dom_get_master(dom, "master5"), 3 ) ) def test_parse_cman_quorum_info(self): parsed = utils.parse_cman_quorum_info("""\ Version: 6.2.0 Config Version: 23 Cluster Name: cluster66 Cluster Id: 22265 Cluster Member: Yes Cluster Generation: 3612 Membership state: Cluster-Member Nodes: 3 Expected votes: 3 Total votes: 3 Node votes: 1 Quorum: 2 Active subsystems: 8 Flags: Ports Bound: 0 Node name: rh66-node2 Node ID: 2 Multicast addresses: 239.192.86.80 Node addresses: 192.168.122.61 ---Votes--- 1 M 3 rh66-node1 2 M 2 rh66-node2 3 M 1 rh66-node3 """) self.assertEquals(True, parsed["quorate"]) self.assertEquals(2, parsed["quorum"]) self.assertEquals( [ {"name": "rh66-node1", "votes": 3, "local": False}, {"name": "rh66-node2", "votes": 2, "local": True}, {"name": "rh66-node3", "votes": 1, "local": False}, ], parsed["node_list"] ) parsed = utils.parse_cman_quorum_info("""\ Version: 6.2.0 Config Version: 23 Cluster Name: cluster66 Cluster Id: 22265 Cluster Member: Yes Cluster Generation: 3612 Membership state: Cluster-Member Nodes: 3 Expected votes: 3 Total votes: 3 Node votes: 1 Quorum: 2 Activity blocked Active subsystems: 8 Flags: Ports Bound: 0 Node name: rh66-node1 Node ID: 1 Multicast addresses: 239.192.86.80 Node addresses: 192.168.122.61 ---Votes--- 1 M 3 rh66-node1 2 X 2 rh66-node2 3 X 1 rh66-node3 """) self.assertEquals(False, parsed["quorate"]) self.assertEquals(2, parsed["quorum"]) self.assertEquals( [ {"name": "rh66-node1", "votes": 3, "local": True}, ], parsed["node_list"] ) parsed = utils.parse_cman_quorum_info("") self.assertEquals(None, parsed) parsed = utils.parse_cman_quorum_info("""\ Version: 6.2.0 Config Version: 23 Cluster Name: cluster66 Cluster Id: 22265 Cluster Member: Yes Cluster Generation: 3612 Membership state: Cluster-Member Nodes: 3 Expected votes: 3 Total votes: 3 Node votes: 1 Quorum: Active subsystems: 8 Flags: Ports Bound: 0 Node name: rh66-node2 Node ID: 2 Multicast addresses: 239.192.86.80 Node addresses: 192.168.122.61 ---Votes--- 1 M 3 rh66-node1 2 M 2 rh66-node2 3 M 1 rh66-node3 """) self.assertEquals(None, parsed) parsed = utils.parse_cman_quorum_info("""\ Version: 6.2.0 Config Version: 23 Cluster Name: cluster66 Cluster Id: 22265 Cluster Member: Yes Cluster Generation: 3612 Membership state: Cluster-Member Nodes: 3 Expected votes: 3 Total votes: 3 Node votes: 1 Quorum: Foo Active subsystems: 8 Flags: Ports Bound: 0 Node name: rh66-node2 Node ID: 2 Multicast addresses: 239.192.86.80 Node addresses: 192.168.122.61 ---Votes--- 1 M 3 rh66-node1 2 M 2 rh66-node2 3 M 1 rh66-node3 """) self.assertEquals(None, parsed) parsed = utils.parse_cman_quorum_info("""\ Version: 6.2.0 Config Version: 23 Cluster Name: cluster66 Cluster Id: 22265 Cluster Member: Yes Cluster Generation: 3612 Membership state: Cluster-Member Nodes: 3 Expected votes: 3 Total votes: 3 Node votes: 1 Quorum: 4 Active subsystems: 8 Flags: Ports Bound: 0 Node name: rh66-node2 Node ID: 2 Multicast addresses: 239.192.86.80 Node addresses: 192.168.122.61 ---Votes--- 1 M 3 rh66-node1 2 M Foo rh66-node2 3 M 1 rh66-node3 """) self.assertEquals(None, parsed) def test_parse_quorumtool_output(self): parsed = utils.parse_quorumtool_output("""\ Quorum information ------------------ Date: Fri Jan 16 13:03:28 2015 Quorum provider: corosync_votequorum Nodes: 3 Node ID: 1 Ring ID: 19860 Quorate: Yes Votequorum information ---------------------- Expected votes: 3 Highest expected: 3 Total votes: 3 Quorum: 2 Flags: Quorate Membership information ---------------------- Nodeid Votes Qdevice Name 1 3 NR rh70-node1 2 2 NR rh70-node2 (local) 3 1 NR rh70-node3 """) self.assertEquals(True, parsed["quorate"]) self.assertEquals(2, parsed["quorum"]) self.assertEquals( [ {"name": "rh70-node1", "votes": 3, "local": False}, {"name": "rh70-node2", "votes": 2, "local": True}, {"name": "rh70-node3", "votes": 1, "local": False}, ], parsed["node_list"] ) parsed = utils.parse_quorumtool_output("""\ Quorum information ------------------ Date: Fri Jan 16 13:03:35 2015 Quorum provider: corosync_votequorum Nodes: 1 Node ID: 1 Ring ID: 19868 Quorate: No Votequorum information ---------------------- Expected votes: 3 Highest expected: 3 Total votes: 1 Quorum: 2 Activity blocked Flags: Membership information ---------------------- Nodeid Votes Qdevice Name 1 1 NR rh70-node1 (local) """) self.assertEquals(False, parsed["quorate"]) self.assertEquals(2, parsed["quorum"]) self.assertEquals( [ {"name": "rh70-node1", "votes": 1, "local": True}, ], parsed["node_list"] ) parsed = utils.parse_quorumtool_output("") self.assertEquals(None, parsed) parsed = utils.parse_quorumtool_output("""\ Quorum information ------------------ Date: Fri Jan 16 13:03:28 2015 Quorum provider: corosync_votequorum Nodes: 3 Node ID: 1 Ring ID: 19860 Quorate: Yes Votequorum information ---------------------- Expected votes: 3 Highest expected: 3 Total votes: 3 Quorum: Flags: Quorate Membership information ---------------------- Nodeid Votes Qdevice Name 1 1 NR rh70-node1 (local) 2 1 NR rh70-node2 3 1 NR rh70-node3 """) self.assertEquals(None, parsed) parsed = utils.parse_quorumtool_output("""\ Quorum information ------------------ Date: Fri Jan 16 13:03:28 2015 Quorum provider: corosync_votequorum Nodes: 3 Node ID: 1 Ring ID: 19860 Quorate: Yes Votequorum information ---------------------- Expected votes: 3 Highest expected: 3 Total votes: 3 Quorum: Foo Flags: Quorate Membership information ---------------------- Nodeid Votes Qdevice Name 1 1 NR rh70-node1 (local) 2 1 NR rh70-node2 3 1 NR rh70-node3 """) self.assertEquals(None, parsed) parsed = utils.parse_quorumtool_output("""\ Quorum information ------------------ Date: Fri Jan 16 13:03:28 2015 Quorum provider: corosync_votequorum Nodes: 3 Node ID: 1 Ring ID: 19860 Quorate: Yes Votequorum information ---------------------- Expected votes: 3 Highest expected: 3 Total votes: 3 Quorum: 2 Flags: Quorate Membership information ---------------------- Nodeid Votes Qdevice Name 1 1 NR rh70-node1 (local) 2 foo NR rh70-node2 3 1 NR rh70-node3 """) self.assertEquals(None, parsed) def test_is_node_stop_cause_quorum_loss(self): quorum_info = { "quorate": False, } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 1, "node_list": [ {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 1, "node_list": [ {"name": "rh70-node3", "votes": 1, "local": True}, ], } self.assertEquals( True, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": False}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": True}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": False}, {"name": "rh70-node2", "votes": 2, "local": True}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": True}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( True, utils.is_node_stop_cause_quorum_loss(quorum_info, True) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": True}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss( quorum_info, False, ["rh70-node3"] ) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": True}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss( quorum_info, False, ["rh70-node2"] ) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": True}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( True, utils.is_node_stop_cause_quorum_loss( quorum_info, False, ["rh70-node1"] ) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 4, "local": True}, {"name": "rh70-node2", "votes": 1, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( False, utils.is_node_stop_cause_quorum_loss( quorum_info, False, ["rh70-node2", "rh70-node3"] ) ) quorum_info = { "quorate": True, "quorum": 4, "node_list": [ {"name": "rh70-node1", "votes": 3, "local": True}, {"name": "rh70-node2", "votes": 2, "local": False}, {"name": "rh70-node3", "votes": 1, "local": False}, ], } self.assertEquals( True, utils.is_node_stop_cause_quorum_loss( quorum_info, False, ["rh70-node2", "rh70-node3"] ) ) def assert_element_id(self, node, node_id): self.assertTrue( isinstance(node, xml.dom.minidom.Element), "element with id '%s' not found" % node_id ) self.assertEquals(node.getAttribute("id"), node_id) if __name__ == "__main__": unittest.main()
gpl-2.0
8,559,471,900,549,907,000
33.059337
87
0.508745
false
NiloFreitas/Deep-Reinforcement-Learning
tensorblock/recipe/recipe_input.py
1
3410
import tensorflow as tf import tensorblock as tb import numpy as np class recipe_input: ####### Add Input def addInput( self , **args ): pars = { **self.defs_input , **args } pars['name'] = self.add_label( self.inputs , 'Input' , pars['name'] , add_order = True ) pars = self.parse_input_pars( pars ) if pars['share'] is not None: self.inputs.append( [ self.node( pars['share'] ) , pars ] ) else: if pars['tensor'] is None: with tf.variable_scope( self.folder + pars['name'] , reuse = False ): self.inputs.append( [ tb.vars.placeholder( shape = pars['shape'] , dtype = pars['dtype'] ) , pars ] ) else: self.inputs.append( [ pars['tensor'] , pars ] ) self.curr_input = pars['name'] return self.inputs[-1][0] ####### Add Variable def addVariable( self , **args ): pars = { **self.defs_variable , **args } pars['name'] = self.add_label( self.variables , 'Variable' , pars['name'] , add_order = True ) pars = self.parse_input_pars( pars ) if pars['share'] is not None: self.variables.append( [ self.node( pars['share'] ) , pars ] ) else: if pars['tensor'] is None: with tf.variable_scope( self.folder + pars['name'] , reuse = False ): self.variables.append( [ pars['type']( pars['shape'] , pars ) , pars ] ) else: if callable( pars['tensor'] ): with tf.variable_scope( self.folder + pars['name'] , reuse = False ): self.variables.append( [ pars['tensor']( pars['shape'] , pars ) , pars ] ) else: if isinstance( pars['tensor'] , np.ndarray ): self.variables.append( [ tb.vars.numpy( pars['tensor'] , pars ) , pars ] ) else: self.variables.append( [ pars['tensor'] , pars ] ) return self.variables[-1][0] ####### Parse Pars def parse_input_pars( self , pars ): if pars['tensor'] is not None: pars['first_none'] = False if isinstance( pars['tensor'] , np.ndarray ): pars['shape'] = pars['tensor'].shape else: pars['shape'] = tb.aux.tf_shape( pars['tensor'] ) if pars['copy'] is not None: # Copying pars['type'] = tb.vars.copy pars['shape'] = self.node( pars['copy'] ) copy_pars = self.pars( pars['copy'] ) pars['out_sides'] = copy_pars['out_sides'] pars['out_channels'] = copy_pars['out_channels'] else: # Nothing pars['shape'] = list( pars['shape'] ) if pars['first_none'] and len( pars['shape'] ) > 1: pars['shape'][0] = None shape = pars['shape'] if pars['out_sides'] is None: if len( shape ) == 2: pars['out_sides'] = shape[1:2] ; if len( shape ) == 4: pars['out_sides'] = shape[1:3] ; if len( shape ) == 5: pars['out_sides'] = shape[1:4] ; if pars['out_channels'] is None: if len( shape ) == 2: pars['out_channels'] = 1 else: pars['out_channels'] = shape[-1] return pars
mit
-6,245,500,325,249,844,000
36.888889
98
0.482405
false
DmitryFillo/berserker_resolver
setup.py
1
1337
import os, sys from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() def dep_dnspython(): name = 'dnspython' if sys.version_info[0] == 3: name += '3' return name setup( name='berserker_resolver', version='2.0.1', author='Dmitry Fillo', author_email='[email protected]', maintainer_email='[email protected]', description=('Fast mass dns resolver which can bypass loadbalancers'), keywords='dns resolver berserker loadbalancer', license='BSD', url='https://github.com/DmitryFillo/berserker_resolver', packages=['berserker_resolver'], install_requires=[dep_dnspython()], tests_require=['mock'], test_suite='tests.get_suite', zip_safe=False, platforms='any', long_description=read('README.rst'), classifiers=[ 'Development Status :: 5 - Production/Stable', 'Topic :: Internet :: Name Service (DNS)', 'Topic :: Utilities', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
bsd-2-clause
917,970,695,018,627,100
30.833333
74
0.618549
false
larsks/cobbler-larsks
cobbler/api.py
1
36927
""" python API module for Cobbler see source for cobbler.py, or pydoc, for example usage. CLI apps and daemons should import api.py, and no other cobbler code. Copyright 2006-2009, Red Hat, Inc Michael DeHaan <[email protected]> This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA """ import yaml import config import utils import action_sync import action_check import action_import import action_reposync import action_status import action_validate import action_buildiso import action_replicate import action_acl import action_report import action_power import action_log import action_hardlink import action_dlcontent from cexceptions import * try: import subprocess as sub_process except: import sub_process import module_loader import kickgen import yumgen import pxegen from utils import _ import logging import time import random import os import xmlrpclib import traceback import exceptions import clogger import item_distro import item_profile import item_system import item_repo import item_image import item_mgmtclass import item_package import item_file ERROR = 100 INFO = 10 DEBUG = 5 # notes on locking: # BootAPI is a singleton object # the XMLRPC variants allow 1 simultaneous request # therefore we flock on /etc/cobbler/settings for now # on a request by request basis. class BootAPI: __shared_state = {} __has_loaded = False # =========================================================== def __init__(self, is_cobblerd=False): """ Constructor """ # FIXME: this should be switchable through some simple system self.__dict__ = BootAPI.__shared_state self.perms_ok = False if not BootAPI.__has_loaded: if os.path.exists("/etc/cobbler/use.couch"): self.use_couch = True else: self.use_couch = False # NOTE: we do not log all API actions, because # a simple CLI invocation may call adds and such # to load the config, which would just fill up # the logs, so we'll do that logging at CLI # level (and remote.py web service level) instead. random.seed() self.is_cobblerd = is_cobblerd try: self.logger = clogger.Logger("/var/log/cobbler/cobbler.log") except CX: # return to CLI/other but perms are not valid # perms_ok is False return # FIMXE: conslidate into 1 server instance self.selinux_enabled = utils.is_selinux_enabled() self.dist = utils.check_dist() self.os_version = utils.os_release() BootAPI.__has_loaded = True module_loader.load_modules() self._config = config.Config(self) self.deserialize() self.authn = self.get_module_from_file( "authentication", "module", "authn_configfile" ) self.authz = self.get_module_from_file( "authorization", "module", "authz_allowall" ) # FIXME: pass more loggers around, and also see that those # using things via tasks construct their own kickgen/yumgen/ # pxegen versus reusing this one, which has the wrong logger # (most likely) for background tasks. self.kickgen = kickgen.KickGen(self._config) self.yumgen = yumgen.YumGen(self._config) self.pxegen = pxegen.PXEGen(self._config, logger=self.logger) self.logger.debug("API handle initialized") self.perms_ok = True # ========================================================== def is_selinux_enabled(self): """ Returns whether selinux is enabled on the cobbler server. We check this just once at cobbler API init time, because a restart is required to change this; this does /not/ check enforce/permissive, nor does it need to. """ return self.selinux_enabled def is_selinux_supported(self): """ Returns whether or not the OS is sufficient enough to run with SELinux enabled (currently EL 5 or later). """ self.dist if self.dist == "redhat" and self.os_version < 5: # doesn't support public_content_t return False return True # ========================================================== def last_modified_time(self): """ Returns the time of the last modification to cobbler, made by any API instance, regardless of the serializer type. """ if not os.path.exists("/var/lib/cobbler/.mtime"): old = os.umask(0x777) fd = open("/var/lib/cobbler/.mtime","w") fd.write("0") fd.close() os.umask(old) return 0 fd = open("/var/lib/cobbler/.mtime") data = fd.read().strip() return float(data) # ========================================================== def log(self,msg,args=None,debug=False): if debug: logger = self.logger.debug else: logger = self.logger.info if args is None: logger("%s" % msg) else: logger("%s; %s" % (msg, str(args))) # ========================================================== def version(self, extended=False): """ What version is cobbler? If extended == False, returns a float for backwards compatibility If extended == True, returns a dict: gitstamp -- the last git commit hash gitdate -- the last git commit date on the builder machine builddate -- the time of the build version -- something like "1.3.2" version_tuple -- something like [ 1, 3, 2 ] """ fd = open("/etc/cobbler/version") ydata = fd.read() fd.close() data = yaml.load(ydata) if not extended: # for backwards compatibility and use with koan's comparisons elems = data["version_tuple"] return int(elems[0]) + 0.1*int(elems[1]) + 0.001*int(elems[2]) else: return data # ========================================================== def clear(self): """ Forget about current list of profiles, distros, and systems # FIXME: is this used anymore? """ return self._config.clear() def __cmp(self,a,b): return cmp(a.name,b.name) # ========================================================== def get_item(self, what, name): self.log("get_item",[what,name],debug=True) return self._config.get_items(what).get(name) # ============================================================= def get_items(self, what): self.log("get_items",[what],debug=True) return self._config.get_items(what) def distros(self): """ Return the current list of distributions """ return self.get_items("distro") def profiles(self): """ Return the current list of profiles """ return self.get_items("profile") def systems(self): """ Return the current list of systems """ return self.get_items("system") def repos(self): """ Return the current list of repos """ return self.get_items("repo") def images(self): """ Return the current list of images """ return self.get_items("image") def settings(self): """ Return the application configuration """ return self._config.settings() def mgmtclasses(self): """ Return the current list of mgmtclasses """ return self.get_items("mgmtclass") def packages(self): """ Return the current list of packages """ return self.get_items("package") def files(self): """ Return the current list of files """ return self.get_items("file") # ======================================================================= def update(self): """ This can be called is no longer used by cobbler. And is here to just avoid breaking older scripts. """ return True # ======================================================================== def copy_item(self, what, ref, newname, logger=None): self.log("copy_item(%s)"%what,[ref.name, newname]) return self.get_items(what).copy(ref,newname,logger=logger) def copy_distro(self, ref, newname): return self.copy_item("distro", ref, newname, logger=None) def copy_profile(self, ref, newname): return self.copy_item("profile", ref, newname, logger=None) def copy_system(self, ref, newname): return self.copy_item("system", ref, newname, logger=None) def copy_repo(self, ref, newname): return self.copy_item("repo", ref, newname, logger=None) def copy_image(self, ref, newname): return self.copy_item("image", ref, newname, logger=None) def copy_mgmtclass(self, ref, newname): return self.copy_item("mgmtclass", ref, newname, logger=None) def copy_package(self, ref, newname): return self.copy_item("package", ref, newname, logger=None) def copy_file(self, ref, newname): return self.copy_item("file", ref, newname, logger=None) # ========================================================================== def remove_item(self, what, ref, recursive=False, delete=True, with_triggers=True, logger=None): if isinstance(what, basestring): if isinstance(ref, basestring): ref = self.get_item(what, ref) if ref is None: return # nothing to remove self.log("remove_item(%s)" % what, [ref.name]) return self.get_items(what).remove(ref.name, recursive=recursive, with_delete=delete, with_triggers=with_triggers, logger=logger) def remove_distro(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("distro", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_profile(self,ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("profile", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_system(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("system", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_repo(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("repo", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_image(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("image", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_mgmtclass(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("mgmtclass", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_package(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("package", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) def remove_file(self, ref, recursive=False, delete=True, with_triggers=True, logger=None): return self.remove_item("file", ref, recursive=recursive, delete=delete, with_triggers=with_triggers, logger=logger) # ========================================================================== def rename_item(self, what, ref, newname, logger=None): self.log("rename_item(%s)"%what,[ref.name,newname]) return self.get_items(what).rename(ref,newname,logger=logger) def rename_distro(self, ref, newname, logger=None): return self.rename_item("distro", ref, newname, logger=logger) def rename_profile(self, ref, newname, logger=None): return self.rename_item("profile", ref, newname, logger=logger) def rename_system(self, ref, newname, logger=None): return self.rename_item("system", ref, newname, logger=logger) def rename_repo(self, ref, newname, logger=None): return self.rename_item("repo", ref, newname, logger=logger) def rename_image(self, ref, newname, logger=None): return self.rename_item("image", ref, newname, logger=logger) def rename_mgmtclass(self, ref, newname, logger=None): return self.rename_item("mgmtclass", ref, newname, logger=logger) def rename_package(self, ref, newname, logger=None): return self.rename_item("package", ref, newname, logger=logger) def rename_file(self, ref, newname, logger=None): return self.rename_item("file", ref, newname, logger=logger) # ========================================================================== # FIXME: add a new_item method def new_distro(self,is_subobject=False): self.log("new_distro",[is_subobject]) return self._config.new_distro(is_subobject=is_subobject) def new_profile(self,is_subobject=False): self.log("new_profile",[is_subobject]) return self._config.new_profile(is_subobject=is_subobject) def new_system(self,is_subobject=False): self.log("new_system",[is_subobject]) return self._config.new_system(is_subobject=is_subobject) def new_repo(self,is_subobject=False): self.log("new_repo",[is_subobject]) return self._config.new_repo(is_subobject=is_subobject) def new_image(self,is_subobject=False): self.log("new_image",[is_subobject]) return self._config.new_image(is_subobject=is_subobject) def new_mgmtclass(self,is_subobject=False): self.log("new_mgmtclass",[is_subobject]) return self._config.new_mgmtclass(is_subobject=is_subobject) def new_package(self,is_subobject=False): self.log("new_package",[is_subobject]) return self._config.new_package(is_subobject=is_subobject) def new_file(self,is_subobject=False): self.log("new_file",[is_subobject]) return self._config.new_file(is_subobject=is_subobject) # ========================================================================== def add_item(self, what, ref, check_for_duplicate_names=False, save=True,logger=None): self.log("add_item(%s)"%what,[ref.name]) return self.get_items(what).add(ref,check_for_duplicate_names=check_for_duplicate_names,save=save,logger=logger) def add_distro(self, ref, check_for_duplicate_names=False, save=True, logger=None): return self.add_item("distro", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_profile(self, ref, check_for_duplicate_names=False,save=True, logger=None): return self.add_item("profile", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_system(self, ref, check_for_duplicate_names=False, check_for_duplicate_netinfo=False, save=True, logger=None): return self.add_item("system", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_repo(self, ref, check_for_duplicate_names=False,save=True,logger=None): return self.add_item("repo", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_image(self, ref, check_for_duplicate_names=False,save=True, logger=None): return self.add_item("image", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_mgmtclass(self, ref, check_for_duplicate_names=False,save=True, logger=None): return self.add_item("mgmtclass", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_package(self, ref, check_for_duplicate_names=False,save=True, logger=None): return self.add_item("package", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) def add_file(self, ref, check_for_duplicate_names=False,save=True, logger=None): return self.add_item("file", ref, check_for_duplicate_names=check_for_duplicate_names, save=save,logger=logger) # ========================================================================== # FIXME: find_items should take all the arguments the other find # methods do. def find_items(self, what, criteria=None): self.log("find_items",[what]) # defaults if criteria is None: criteria={} items=self._config.get_items(what) # empty criteria returns everything if criteria == {}: res=items else: res=items.find(return_list=True, no_errors=False, **criteria) return res def find_distro(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.distros().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_profile(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.profiles().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_system(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.systems().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_repo(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.repos().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_image(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.images().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_mgmtclass(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.mgmtclasses().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_package(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.packages().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) def find_file(self, name=None, return_list=False, no_errors=False, **kargs): return self._config.files().find(name=name, return_list=return_list, no_errors=no_errors, **kargs) # ========================================================================== def __since(self,mtime,collector,collapse=False): """ Called by get_*_since functions. """ results1 = collector() results2 = [] for x in results1: if x.mtime == 0 or x.mtime >= mtime: if not collapse: results2.append(x) else: results2.append(x.to_datastruct()) return results2 def get_distros_since(self,mtime,collapse=False): """ Returns distros modified since a certain time (in seconds since Epoch) collapse=True specifies returning a hash instead of objects. """ return self.__since(mtime,self.distros,collapse=collapse) def get_profiles_since(self,mtime,collapse=False): return self.__since(mtime,self.profiles,collapse=collapse) def get_systems_since(self,mtime,collapse=False): return self.__since(mtime,self.systems,collapse=collapse) def get_repos_since(self,mtime,collapse=False): return self.__since(mtime,self.repos,collapse=collapse) def get_images_since(self,mtime,collapse=False): return self.__since(mtime,self.images,collapse=collapse) def get_mgmtclasses_since(self,mtime,collapse=False): return self.__since(mtime,self.mgmtclasses,collapse=collapse) def get_packages_since(self,mtime,collapse=False): return self.__since(mtime,self.packages,collapse=collapse) def get_files_since(self,mtime,collapse=False): return self.__since(mtime,self.files,collapse=collapse) # ========================================================================== def dump_vars(self, obj, format=False): return obj.dump_vars(format) # ========================================================================== def auto_add_repos(self): """ Import any repos this server knows about and mirror them. Credit: Seth Vidal. """ self.log("auto_add_repos") try: import yum except: raise CX(_("yum is not installed")) version = yum.__version__ (a,b,c) = version.split(".") version = a* 1000 + b*100 + c if version < 324: raise CX(_("need yum > 3.2.4 to proceed")) base = yum.YumBase() base.doRepoSetup() repos = base.repos.listEnabled() if len(repos) == 0: raise CX(_("no repos enabled/available -- giving up.")) for repo in repos: url = repo.urls[0] cobbler_repo = self.new_repo() auto_name = repo.name.replace(" ","") # FIXME: probably doesn't work for yum-rhn-plugin ATM cobbler_repo.set_mirror(url) cobbler_repo.set_name(auto_name) print "auto adding: %s (%s)" % (auto_name, url) self._config.repos().add(cobbler_repo,save=True) # run cobbler reposync to apply changes return True # ========================================================================== def get_repo_config_for_profile(self,obj): return self.yumgen.get_yum_config(obj,True) def get_repo_config_for_system(self,obj): return self.yumgen.get_yum_config(obj,False) # ========================================================================== def get_template_file_for_profile(self,obj,path): template_results = self.pxegen.write_templates(obj,False,path) if template_results.has_key(path): return template_results[path] else: return "# template path not found for specified profile" def get_template_file_for_system(self,obj,path): template_results = self.pxegen.write_templates(obj,False,path) if template_results.has_key(path): return template_results[path] else: return "# template path not found for specified system" # ========================================================================== def generate_kickstart(self,profile,system): self.log("generate_kickstart") if system: return self.kickgen.generate_kickstart_for_system(system) else: return self.kickgen.generate_kickstart_for_profile(profile) # ========================================================================== def check(self, logger=None): """ See if all preqs for network booting are valid. This returns a list of strings containing instructions on things to correct. An empty list means there is nothing to correct, but that still doesn't mean there are configuration errors. This is mainly useful for human admins, who may, for instance, forget to properly set up their TFTP servers for PXE, etc. """ self.log("check") check = action_check.BootCheck(self._config, logger=logger) return check.run() # ========================================================================== def dlcontent(self,force=False,logger=None): """ Downloads bootloader content that may not be avialable in packages for the given arch, ex: if installing on PPC, get syslinux. If installing on x86_64, get elilo, etc. """ # FIXME: teach code that copies it to grab from the right place self.log("dlcontent") grabber = action_dlcontent.ContentDownloader(self._config, logger=logger) return grabber.run(force) # ========================================================================== def validateks(self, logger=None): """ Use ksvalidator (from pykickstart, if available) to determine whether the cobbler kickstarts are going to be (likely) well accepted by Anaconda. Presence of an error does not indicate the kickstart is bad, only that the possibility exists. ksvalidator is not available on all platforms and can not detect "future" kickstart format correctness. """ self.log("validateks") validator = action_validate.Validate(self._config, logger=logger) return validator.run() # ========================================================================== def sync(self,verbose=False, logger=None): """ Take the values currently written to the configuration files in /etc, and /var, and build out the information tree found in /tftpboot. Any operations done in the API that have not been saved with serialize() will NOT be synchronized with this command. """ self.log("sync") sync = self.get_sync(verbose=verbose, logger=logger) return sync.run() # ========================================================================== def get_sync(self,verbose=False,logger=None): self.dhcp = self.get_module_from_file( "dhcp", "module", "manage_isc" ).get_manager(self._config,logger) self.dns = self.get_module_from_file( "dns", "module", "manage_bind" ).get_manager(self._config,logger) self.tftpd = self.get_module_from_file( "tftpd", "module", "in_tftpd", ).get_manager(self._config,logger) return action_sync.BootSync(self._config,dhcp=self.dhcp,dns=self.dns,tftpd=self.tftpd,verbose=verbose,logger=logger) # ========================================================================== def reposync(self, name=None, tries=1, nofail=False, logger=None): """ Take the contents of /var/lib/cobbler/repos and update them -- or create the initial copy if no contents exist yet. """ self.log("reposync",[name]) reposync = action_reposync.RepoSync(self._config, tries=tries, nofail=nofail, logger=logger) return reposync.run(name) # ========================================================================== def status(self,mode,logger=None): statusifier = action_status.BootStatusReport(self._config,mode,logger=logger) return statusifier.run() # ========================================================================== def import_tree(self,mirror_url,mirror_name,network_root=None,kickstart_file=None,rsync_flags=None,arch=None,breed=None,os_version=None,logger=None): """ Automatically import a directory tree full of distribution files. mirror_url can be a string that represents a path, a user@host syntax for SSH, or an rsync:// address. If mirror_url is a filesystem path and mirroring is not desired, set network_root to something like "nfs://path/to/mirror_url/root" """ self.log("import_tree",[mirror_url, mirror_name, network_root, kickstart_file, rsync_flags]) importer_modules = self.get_modules_in_category("manage/import") for importer_module in importer_modules: manager = importer_module.get_import_manager(self._config,logger) if 1:#try: (found,pkgdir) = manager.check_for_signature(mirror_url,breed) if found: self.log("running import manager: %s" % manager.what()) return manager.run(pkgdir,mirror_url,mirror_name,network_root,kickstart_file,rsync_flags,arch,breed,os_version) #except: # self.log("an error occured while running the import manager") # continue self.log("No import managers found a valid signature at the location specified") return False # ========================================================================== def acl_config(self,adduser=None,addgroup=None,removeuser=None,removegroup=None, logger=None): """ Configures users/groups to run the cobbler CLI as non-root. Pass in only one option at a time. Powers "cobbler aclconfig" """ acl = action_acl.AclConfig(self._config, logger) return acl.run( adduser=adduser, addgroup=addgroup, removeuser=removeuser, removegroup=removegroup ) # ========================================================================== def serialize(self): """ Save the config file(s) to disk. Cobbler internal use only. """ return self._config.serialize() def deserialize(self): """ Load the current configuration from config file(s) Cobbler internal use only. """ return self._config.deserialize() def deserialize_raw(self,collection_name): """ Get the collection back just as raw data. Cobbler internal use only. """ return self._config.deserialize_raw(collection_name) def deserialize_item_raw(self,collection_name,obj_name): """ Get an object back as raw data. Can be very fast for shelve or catalog serializers Cobbler internal use only. """ return self._config.deserialize_item_raw(collection_name,obj_name) # ========================================================================== def get_module_by_name(self,module_name): """ Returns a loaded cobbler module named 'name', if one exists, else None. Cobbler internal use only. """ return module_loader.get_module_by_name(module_name) def get_module_from_file(self,section,name,fallback=None): """ Looks in /etc/cobbler/modules.conf for a section called 'section' and a key called 'name', and then returns the module that corresponds to the value of that key. Cobbler internal use only. """ return module_loader.get_module_from_file(section,name,fallback) def get_modules_in_category(self,category): """ Returns all modules in a given category, for instance "serializer", or "cli". Cobbler internal use only. """ return module_loader.get_modules_in_category(category) # ========================================================================== def authenticate(self,user,password): """ (Remote) access control. Cobbler internal use only. """ rc = self.authn.authenticate(self,user,password) self.log("authenticate",[user,rc]) return rc def authorize(self,user,resource,arg1=None,arg2=None): """ (Remote) access control. Cobbler internal use only. """ rc = self.authz.authorize(self,user,resource,arg1,arg2) self.log("authorize",[user,resource,arg1,arg2,rc],debug=True) return rc # ========================================================================== def build_iso(self,iso=None,profiles=None,systems=None,buildisodir=None,distro=None,standalone=None,source=None, exclude_dns=None, logger=None): builder = action_buildiso.BuildIso(self._config, logger=logger) return builder.run( iso=iso, profiles=profiles, systems=systems, buildisodir=buildisodir, distro=distro, standalone=standalone, source=source, exclude_dns=exclude_dns ) # ========================================================================== def hardlink(self, logger=None): linker = action_hardlink.HardLinker(self._config, logger=logger) return linker.run() # ========================================================================== def replicate(self, cobbler_master = None, distro_patterns="", profile_patterns="", system_patterns="", repo_patterns="", image_patterns="", mgmtclass_patterns=None, package_patterns=None, file_patterns=None, prune=False, omit_data=False, sync_all=False, logger=None): """ Pull down data/configs from a remote cobbler server that is a master to this server. """ replicator = action_replicate.Replicate(self._config, logger=logger) return replicator.run( cobbler_master = cobbler_master, distro_patterns = distro_patterns, profile_patterns = profile_patterns, system_patterns = system_patterns, repo_patterns = repo_patterns, image_patterns = image_patterns, mgmtclass_patterns = mgmtclass_patterns, package_patterns = package_patterns, file_patterns = file_patterns, prune = prune, omit_data = omit_data, sync_all = sync_all ) # ========================================================================== def report(self, report_what = None, report_name = None, report_type = None, report_fields = None, report_noheaders = None): """ Report functionality for cobbler """ reporter = action_report.Report(self._config) return reporter.run(report_what = report_what, report_name = report_name,\ report_type = report_type, report_fields = report_fields,\ report_noheaders = report_noheaders) # ========================================================================== def get_kickstart_templates(self): return utils.get_kickstar_templates(self) # ========================================================================== def power_on(self, system, user=None, password=None, logger=None): """ Powers up a system that has power management configured. """ return action_power.PowerTool(self._config,system,self,user,password,logger=logger).power("on") def power_off(self, system, user=None, password=None, logger=None): """ Powers down a system that has power management configured. """ return action_power.PowerTool(self._config,system,self,user,password,logger=logger).power("off") def reboot(self,system, user=None, password=None, logger=None): """ Cycles power on a system that has power management configured. """ self.power_off(system, user, password, logger=logger) time.sleep(5) return self.power_on(system, user, password, logger=logger) def power_status(self, system, user=None, password=None, logger=None): """ Returns the power status for a system that has power management configured. @return: 0 the system is powered on, False if it's not or None on error """ return action_power.PowerTool(self._config, system, self, user, password, logger = logger).power("status") # ========================================================================== def clear_logs(self, system, logger=None): """ Clears console and anamon logs for system """ return action_log.LogTool(self._config,system,self, logger=logger).clear() def get_os_details(self): return (self.dist, self.os_version)
gpl-2.0
7,651,953,824,866,149,000
37.993664
157
0.578168
false
Adamssss/projectEuler
Problem 001-150 Python/pb103.py
1
1978
import math import time t1 = time.time() # using all the numbers # the sub pairs cover the exact original set def exactsub(oset): l = len(oset) if l == 2: return [[[oset[0]],[oset[1]]]] result = [] f = oset[0] rest = oset[1:] result.append([[f],rest]) for i in exactsub(rest): a = i[0] b = i[1] result.append([a+[f],b]) result.append([a,b+[f]]) return result def allsub(oset): temp = exactsub(oset) result = temp[:] for i in temp: if len(i[0]) > 1: result += exactsub(i[0]) if len(i[1]) > 1: result += exactsub(i[1]) return result def checksub(setp): B = setp[0] C = setp[1] sb = sum(B) sc = sum(C) if sb == sc: return False lb = len(B) lc = len(C) if lb > lc and sb <= sc: return False if lb < lc and sb >= sc: return False return True def checkset(tset): for i in allsub(tset): if not checksub(i): return False return True def toString(aset): temp = 0 for i in aset: dig = math.floor(math.log10(i)+1) temp = temp*math.pow(10,dig)+i return int(temp) sset = [[],[1],[1,2],[2,3,4],[3,5,6,7],[6,9,11,12,13],[11,18,19,20,22,25]] AL = 2 def near(n): if n == 1: result = [] for i in range(0,AL*2): result.append([i]) return result result = [] for i in range(0,AL*2): for j in near(n-1): result.append([i]+j) return result def addaprox(seta,setb): result = seta[:] for i in range(len(seta)): result[i] += setb[i] return result def makeset(n): temp = sset[n-1] a = temp[n//2-1] base = [a] for i in range(n-1): base.append(temp[i]+a-AL) for j in near(n): temp = addaprox(base,j) if checkset(temp): return temp print(toString(makeset(7))) print("time:",time.time()-t1)
mit
849,634,466,316,029,700
18.979798
74
0.505561
false
NeurodataWithoutBorders/api-python
examples/create_scripts/extensions/e-trajectorySeries2.py
1
2936
# Definitions of extension to TimeSeries to store trajectory information # All format specifications must be stored in dictionary "fs" # "mnts2" is the "namespace" for this extension # This extension explicitly specifies meaning for each column of dataset data {"fs": {"mnts2": { "info": { "name": "Sabes lab data trajectory series", "version": "1.0", "date": "Oct 3, 2016", "author": "Jeff Teeters", "contact": "[email protected]", "description": ("Extension to store timeseries of hand trajectories for Sabes lab data") }, "schema": { "<TrajectorySeries>/": { "merge": ["core:<SpatialSeries>/"], "attributes": { "ancestry": { "data_type": "text", "dimensions": ["3"], "value": ["TimeSeries", "SpatialSeries", "TrajectorySeries"], "const": True}, "help": { "data_type": "text", "value": "Trajectory of hand movement positions", "const": True}, }, "data": { "description": ("Measurements of hand trajectory, recorded at each point of time."), "dimensions": ["num_times", "trajectories"], "data_type": "float32", "trajectories": { "type": "structure", # define components of trajectories dimension "components": [ { "alias": "s1_x", "unit": "meter" }, { "alias": "s1_y", "unit": "meter" }, { "alias": "s1_z", "unit": "meter" }, { "alias": "s1_pitch", "unit": "radian" }, { "alias": "s1_roll", "unit": "radian" }, { "alias": "s1_yaw", "unit": "radian" }, { "alias": "s2_x", "unit": "meter" }, { "alias": "s2_y", "unit": "meter" }, { "alias": "s2_z", "unit": "meter" }, { "alias": "s2_pitch", "unit": "radian" }, { "alias": "s2_roll", "unit": "radian" }, { "alias": "s2_yaw", "unit": "radian" }, { "alias": "s3_x", "unit": "meter" }, { "alias": "s3_y", "unit": "meter" }, { "alias": "s3_z", "unit": "meter" }, { "alias": "s3_pitch", "unit": "radian" }, { "alias": "s3_roll", "unit": "radian" }, { "alias": "s3_yaw", "unit": "radian" }, { "alias": "s4_x", "unit": "meter" }, { "alias": "s4_y", "unit": "meter" }, { "alias": "s4_z", "unit": "meter" }, { "alias": "s4_pitch", "unit": "radian" }, { "alias": "s4_roll", "unit": "radian" }, { "alias": "s4_yaw", "unit": "radian" } ] }, } } } }}}
bsd-3-clause
-9,183,462,637,381,410,000
42.835821
96
0.419619
false
tangyaohua/dl4mt
session2/lm/deepy/trainers/optimize.py
1
4684
#!/usr/bin/env python # -*- coding: utf-8 -*- import logging as loggers import numpy as np import theano import theano.tensor as T from theano.ifelse import ifelse from ..utils import FLOATX, dim_to_var, EPSILON from .util import wrap_core, multiple_l2_norm from ..conf import TrainerConfig logging = loggers.getLogger(__name__) def optimize_updates(params, gradients, config=None, shapes=None): """ General optimization function for Theano. Parameters: params - parameters gradients - gradients config - training config Returns: Theano updates :type config: deepy.TrainerConfig or dict """ if config and isinstance(config, dict): config = TrainerConfig(config) # Clipping if config: clip_value = config.get("gradient_clipping", None) if clip_value: clip_constant = T.constant(clip_value, dtype=FLOATX) if config.avoid_compute_embed_norm: grad_norm = multiple_l2_norm([t[1] for t in zip(params, gradients) if not t[0].name.startswith("W_embed")]) else: grad_norm = multiple_l2_norm(gradients) isnan = T.or_(T.isnan(grad_norm), T.isinf(grad_norm)) multiplier = ifelse(grad_norm < clip_constant, T.constant(1., dtype=FLOATX), clip_constant / (grad_norm + EPSILON)) # Clip clipped_gradients = [] for param, g in zip(params, gradients): g = multiplier * g if config.avoid_nan: g = T.switch(isnan, np.float32(0.1) * param, g) if config.gradient_tolerance: g = ifelse(grad_norm > config.gradient_tolerance, T.zeros_like(g) + EPSILON, g) clipped_gradients.append(g) gradients = clipped_gradients # Regularization if config and config.weight_l2: regularized_gradients = [] for param, grad in zip(params, gradients): grad = grad + (2 * config.weight_l2 * param) regularized_gradients.append(grad) gradients = regularized_gradients # Avoid nan but not computing the norm # This is not recommended if config and config.avoid_nan and not config.gradient_clipping: logging.info("avoid NaN gradients") new_gradients = [] for grad in gradients: new_grad = ifelse(T.isnan(grad).any(), T.zeros_like(grad) + EPSILON, grad) new_gradients.append(new_grad) gradients = new_gradients # Find method method = "SGD" if config: method = config.get("method", method).upper() # Get Function func = None if method in ["SGD", "ADAGRAD", "ADADELTA", "FINETUNING_ADAGRAD"]: from cores.ada_family import ada_family_core func = ada_family_core elif method == "ADAM": from cores.adam import adam_core func = adam_core elif method == "RMSPROP": from cores.rmsprop import rmsprop_core func = rmsprop_core elif method == "MOMENTUM": from cores.momentum import momentum_core func = momentum_core if not func: raise NotImplementedError("method '%s' is not supported" % method) logging.info("optimize method=%s parameters=%s" % (method, str(params))) free_parameters = [] return_vals = wrap_core(func, config, params, gradients) if type(return_vals) == list and type(return_vals[0]) == list: updates, free_parameters = return_vals else: updates = return_vals # No free param recording if config and not config.record_free_params: free_parameters = [] # Weight bound if config.weight_bound: logging.info("apply weight bound of %.2f" % config.weight_bound) new_updates = [] for param, update_value in updates: bounded_value = (update_value * (T.abs_(update_value) <= config.weight_bound) + config.weight_bound * (update_value > config.weight_bound) + -config.weight_bound * (update_value < -config.weight_bound)) new_updates.append((param, bounded_value)) updates = new_updates return updates, free_parameters def optimize_function(params, config=None): """ Create a optimizing function receives gradients. Parameters: params - parameters config - training configuration Returns: updating function receives gradients """ gs = [dim_to_var(p.ndim) for p in params] updates, _ = optimize_updates(params, gs, config) return theano.function(gs, [], updates=updates)
bsd-3-clause
-5,147,573,832,780,105,000
33.696296
123
0.609308
false
TeamHG-Memex/hgprofiler
lib/cli/run_server.py
1
1340
import app import cli class RunServerCli(cli.BaseCli): """ A tool for running a development server. """ def _get_args(self, arg_parser): """ Customize arguments. """ arg_parser.add_argument( '--debug', action='store_true', help='Enable debug mode: errors produce stack traces and' \ ' the server auto reloads on source code changes.' ) arg_parser.add_argument( '--debug-db', action='store_true', help='Print database queries.' ) arg_parser.add_argument( '--ip', default='127.0.0.1', help='Specify an IP address to bind to. (Defaults to loopback.)' ) arg_parser.add_argument( '--latency', type=float, metavar='L', help='Delay each request by <L> seconds.' ) def _run(self, args, config): """ Main entry point. """ flask_app = app.bootstrap( debug=args.debug, debug_db=args.debug_db, latency=args.latency, log_level=args.verbosity ) # Disable secure cookies for the development server. flask_app.config["SESSION_COOKIE_SECURE"] = False flask_app.run(host=args.ip, threaded=True)
apache-2.0
466,433,427,109,874,750
26.346939
76
0.526866
false
yephper/django
django/conf/locale/en_GB/formats.py
1
2159
# -*- encoding: utf-8 -*- # This file is distributed under the same license as the Django package. # from __future__ import unicode_literals # The *_FORMAT strings use the Django date format syntax, # see http://docs.djangoproject.com/en/dev/ref/templates/builtins/#date DATE_FORMAT = 'j M Y' # '25 Oct 2006' TIME_FORMAT = 'P' # '2:30 p.m.' DATETIME_FORMAT = 'j M Y, P' # '25 Oct 2006, 2:30 p.m.' YEAR_MONTH_FORMAT = 'F Y' # 'October 2006' MONTH_DAY_FORMAT = 'j F' # '25 October' SHORT_DATE_FORMAT = 'd/m/Y' # '25/10/2006' SHORT_DATETIME_FORMAT = 'd/m/Y P' # '25/10/2006 2:30 p.m.' FIRST_DAY_OF_WEEK = 1 # Monday # The *_INPUT_FORMATS strings use the Python strftime format syntax, # see http://docs.python.org/library/datetime.html#strftime-strptime-behavior DATE_INPUT_FORMATS = [ '%d/%m/%Y', '%d/%m/%y', # '25/10/2006', '25/10/06' # '%b %d %Y', '%b %d, %Y', # 'Oct 25 2006', 'Oct 25, 2006' # '%d %b %Y', '%d %b, %Y', # '25 Oct 2006', '25 Oct, 2006' # '%B %d %Y', '%B %d, %Y', # 'October 25 2006', 'October 25, 2006' # '%d %B %Y', '%d %B, %Y', # '25 October 2006', '25 October, 2006' ] DATETIME_INPUT_FORMATS = [ '%Y-%m-%d %H:%M:%S', # '2006-10-25 14:30:59' '%Y-%m-%d %H:%M:%S.%f', # '2006-10-25 14:30:59.000200' '%Y-%m-%d %H:%M', # '2006-10-25 14:30' '%Y-%m-%d', # '2006-10-25' '%d/%m/%Y %H:%M:%S', # '25/10/2006 14:30:59' '%d/%m/%Y %H:%M:%S.%f', # '25/10/2006 14:30:59.000200' '%d/%m/%Y %H:%M', # '25/10/2006 14:30' '%d/%m/%Y', # '25/10/2006' '%d/%m/%y %H:%M:%S', # '25/10/06 14:30:59' '%d/%m/%y %H:%M:%S.%f', # '25/10/06 14:30:59.000200' '%d/%m/%y %H:%M', # '25/10/06 14:30' '%d/%m/%y', # '25/10/06' ] DECIMAL_SEPARATOR = '.' THOUSAND_SEPARATOR = ',' NUMBER_GROUPING = 3
bsd-3-clause
-7,802,788,513,267,937,000
49.404762
79
0.443261
false
mcf-rocha/swagger-graph
api-ingestor.py
1
2985
import sys import pysvn import json import configparser, os from pymongo import MongoClient config = configparser.RawConfigParser() config.read('/Users/mcf/configuration.cfg') svnroot = config.get('SVN', 'root') svnpath = config.get('SVN', 'path') svnusername = config.get('SVN', 'username') svnpwd = config.get('SVN', 'pwd') mongohost = config.get('MongoDB', 'host') mongoport = config.getint('MongoDB', 'port') svnclient = pysvn.Client() svnclient.set_default_username(svnusername) svnclient.set_default_password(svnpwd) try: dirlist = svnclient.list(svnroot+svnpath,peg_revision=pysvn.Revision(pysvn.opt_revision_kind.head)) except BaseException as be: print("Could not connect to SVN. Execution aborted.",) print(be.__str__()) sys.exit() else: mongoclient = MongoClient(mongohost, mongoport) print("Cleaning the MongoDB collection...") mongodb = mongoclient.catlog mongodb.swagger.drop() for item, locked in dirlist: if item.kind == pysvn.node_kind.file: #print("Starting to process document {}".format(item.path)) nomearquivoswagger = item.path.rpartition("/")[2] print("Starting to process document {}".format(nomearquivoswagger)) swaggercontentbyte = svnclient.cat(item.path,peg_revision=pysvn.Revision(pysvn.opt_revision_kind.head)) try: swaggercontentstring = swaggercontentbyte.decode("utf-8") except: print(" erro ao decode utf-8. tentando utf-8-sig") sys.exit(1) ##-------------------AJUSTES REQUERIDOS PELO SINTAXE DO SWAGGER QUE NÃO SÃO PERMITIDOS NO MONGO swaggercontentstring2 = swaggercontentstring.replace("$ref","ref") ##------------------- #print(swaggercontentstring2) try: swaggerjson = json.loads(swaggercontentstring2) except json.decoder.JSONDecodeError as jde: swaggercontentstring = swaggercontentbyte.decode("utf-8-sig") swaggercontentstring2 = swaggercontentstring.replace("$ref","ref") swaggerjson = json.loads(swaggercontentstring2) #else: # print(" An encoding problem has happened when inserting Swagger information into the MongoDB collection... {}".format(getSwaggerTitle(swaggerjson))) # sys.exit(1) swaggerjson["_id"] = nomearquivoswagger print(" Inserting Swagger information into the MongoDB collection...") try: result = mongodb.swagger.insert_one(swaggerjson) except BaseException as mongoex: print(" The Swagger information could not be inserted into the MongoDB collection.") print(mongoex.__str__()) print() else: print(" Document _id: {}".format(nomearquivoswagger,result.inserted_id)) mongoclient.close()
mit
6,348,466,525,802,908,000
40.444444
170
0.628227
false
TwoBitAlchemist/NeoAlchemy
neoalchemy/cypher/verbs.py
1
4169
from __future__ import unicode_literals import six from ..shared.objects import Property from .operations import CypherExpression, ComparisonExpression, QueryParams class CypherQuery(list): def __init__(self, graph_obj, use_full_pattern=False): self.params = QueryParams() try: verb = self.verb except AttributeError: verb = self.__class__.__name__.upper() pattern = graph_obj.pattern(inline_props=use_full_pattern) super(CypherQuery, self).__init__(['%s %s' % (verb, pattern)]) def delete(self, *args, **kw): detach = kw.get('detach') keyword = 'DETACH DELETE ' if detach else 'DELETE ' self.append(keyword + ', '.join(arg.var for arg in args)) return self def limit(self, limit): self.append('LIMIT %i' % int(limit)) return self def order_by(self, *args, **kw): stmt = 'ORDER BY ' + ', '.join(arg.var for arg in args) if kw.get('desc'): stmt += ' DESC' self.append(stmt) return self def remove(self, *args): self.append('REMOVE ' + ', '.join(arg.var for arg in args)) return self def return_(self, *args): if args: self.append('RETURN ' + ', '.join(arg.var for arg in args)) else: self.append('RETURN *') return self def set(self, *props): prop_list = [self._add_expression(prop) for prop in props] if prop_list: self.append(' SET ' + ', '.join(prop_list)) return self def skip(self, skip): self.append('SKIP %i' % int(skip)) return self def where(self, *exprs, **kw): or_ = kw.pop('or_', False) stmt_list = [self._add_expression(expr) for expr in exprs] if stmt_list: statements = ' AND '.join(stmt_list) if any(keyword.rjust(9) == self[-1][:9] for keyword in ('WHERE', 'AND', 'OR')): keyword = 'AND ' if not or_ else 'OR ' else: keyword = 'WHERE ' self.append(keyword.rjust(10) + statements) return self def with_(self, *args): self.append('WITH ' + ', '.join(arg.var for arg in args)) return self def _add_expression(self, expr): if isinstance(expr, Property): prop = expr expr = ComparisonExpression(prop, prop.value, '=') else: if not isinstance(expr, CypherExpression): raise ValueError('Must be CypherExpression or Property') for key, value in expr.compile().params.items(): self.params[key] = value if self.params.last_key.startswith('param'): expr.replace(key, self.params.last_key) return str(expr) def __str__(self): return '\n'.join(map(str, self)) def __and__(self, query): self.extend(query) self.params.update(query.params) return self def __or__(self, query): self.append('UNION ALL') self.extend(query) self.params.update(query.params) return self def __xor__(self, query): self.append('UNION') self.extend(query) self.params.update(query.params) return self class Create(CypherQuery): def __init__(self, graph_obj): super(Create, self).__init__(graph_obj) self.set(*graph_obj.values()) class Match(CypherQuery): def __init__(self, graph_obj, optional=False): if optional: self.verb = 'OPTIONAL MATCH' super(Match, self).__init__(graph_obj) self.where(*(v for k, v in graph_obj.items() if k in graph_obj.bound_keys)) class Merge(CypherQuery): def __init__(self, graph_obj): super(Merge, self).__init__(graph_obj, use_full_pattern=True) self.params.update({p.param: p.value for k, p in graph_obj.items() if k in graph_obj.bound_keys}) def on_create(self): self.append('ON CREATE') return self def on_match(self): self.append('ON MATCH') return self
mit
-7,028,930,501,125,680,000
28.992806
75
0.550971
false
mahabs/nitro
nssrc/com/citrix/netscaler/nitro/resource/config/dns/dnspolicy_dnspolicylabel_binding.py
1
6004
# # Copyright (c) 2008-2015 Citrix Systems, Inc. # # Licensed under the Apache License, Version 2.0 (the "License") # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from nssrc.com.citrix.netscaler.nitro.resource.base.base_resource import base_resource from nssrc.com.citrix.netscaler.nitro.resource.base.base_resource import base_response from nssrc.com.citrix.netscaler.nitro.service.options import options from nssrc.com.citrix.netscaler.nitro.exception.nitro_exception import nitro_exception from nssrc.com.citrix.netscaler.nitro.util.nitro_util import nitro_util class dnspolicy_dnspolicylabel_binding(base_resource) : """ Binding class showing the dnspolicylabel that can be bound to dnspolicy. """ def __init__(self) : self._boundto = "" self._priority = 0 self._activepolicy = 0 self._gotopriorityexpression = "" self._labeltype = "" self._labelname = "" self._name = "" self.___count = 0 @property def boundto(self) : """Location where policy is bound. """ try : return self._boundto except Exception as e: raise e @boundto.setter def boundto(self, boundto) : """Location where policy is bound. """ try : self._boundto = boundto except Exception as e: raise e @property def name(self) : """Name of the DNS policy. """ try : return self._name except Exception as e: raise e @name.setter def name(self, name) : """Name of the DNS policy. """ try : self._name = name except Exception as e: raise e @property def priority(self) : """Specifies the priority of the policy. """ try : return self._priority except Exception as e: raise e @property def labelname(self) : """Name of the label to invoke if the current policy rule evaluates to TRUE. """ try : return self._labelname except Exception as e: raise e @property def gotopriorityexpression(self) : """Expression specifying the priority of the next policy which will get evaluated if the current policy rule evaluates to TRUE. """ try : return self._gotopriorityexpression except Exception as e: raise e @property def labeltype(self) : """Type of policy label invocation.<br/>Possible values = reqvserver, resvserver, policylabel. """ try : return self._labeltype except Exception as e: raise e @property def activepolicy(self) : """Indicates whether policy is bound or not. """ try : return self._activepolicy except Exception as e: raise e def _get_nitro_response(self, service, response) : """ converts nitro response into object and returns the object array in case of get request. """ try : result = service.payload_formatter.string_to_resource(dnspolicy_dnspolicylabel_binding_response, response, self.__class__.__name__) if(result.errorcode != 0) : if (result.errorcode == 444) : service.clear_session(self) if result.severity : if (result.severity == "ERROR") : raise nitro_exception(result.errorcode, str(result.message), str(result.severity)) else : raise nitro_exception(result.errorcode, str(result.message), str(result.severity)) return result.dnspolicy_dnspolicylabel_binding except Exception as e : raise e def _get_object_name(self) : """ Returns the value of object identifier argument """ try : if (self.name) : return str(self.name) return None except Exception as e : raise e @classmethod def get(cls, service, name) : """ Use this API to fetch dnspolicy_dnspolicylabel_binding resources. """ try : obj = dnspolicy_dnspolicylabel_binding() obj.name = name response = obj.get_resources(service) return response except Exception as e: raise e @classmethod def get_filtered(cls, service, name, filter_) : """ Use this API to fetch filtered set of dnspolicy_dnspolicylabel_binding resources. Filter string should be in JSON format.eg: "port:80,servicetype:HTTP". """ try : obj = dnspolicy_dnspolicylabel_binding() obj.name = name option_ = options() option_.filter = filter_ response = obj.getfiltered(service, option_) return response except Exception as e: raise e @classmethod def count(cls, service, name) : """ Use this API to count dnspolicy_dnspolicylabel_binding resources configued on NetScaler. """ try : obj = dnspolicy_dnspolicylabel_binding() obj.name = name option_ = options() option_.count = True response = obj.get_resources(service, option_) if response : return response[0].__dict__['___count'] return 0 except Exception as e: raise e @classmethod def count_filtered(cls, service, name, filter_) : """ Use this API to count the filtered set of dnspolicy_dnspolicylabel_binding resources. Filter string should be in JSON format.eg: "port:80,servicetype:HTTP". """ try : obj = dnspolicy_dnspolicylabel_binding() obj.name = name option_ = options() option_.count = True option_.filter = filter_ response = obj.getfiltered(service, option_) if response : return response[0].__dict__['___count'] return 0 except Exception as e: raise e class Labeltype: reqvserver = "reqvserver" resvserver = "resvserver" policylabel = "policylabel" class dnspolicy_dnspolicylabel_binding_response(base_response) : def __init__(self, length=1) : self.dnspolicy_dnspolicylabel_binding = [] self.errorcode = 0 self.message = "" self.severity = "" self.sessionid = "" self.dnspolicy_dnspolicylabel_binding = [dnspolicy_dnspolicylabel_binding() for _ in range(length)]
apache-2.0
-9,121,447,943,516,050,000
26.167421
134
0.698368
false
pudo/regenesis
legacy/regenesis/test_table.py
1
2695
# -*- coding: utf-8 -*- import unittest import table MIKROZENSUS_11111_0001 = """GENESIS-Tabelle: Temporär Gebietsfläche: Bundesländer, Stichtag;;;;;;;;;;;;; Feststellung des Gebietsstands;;;;;;;;;;;;; Gebietsfläche (qkm);;;;;;;;;;;;; ;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag;Stichtag ;31.12.1995;31.12.1996;31.12.1997;31.12.1998;31.12.1999;31.12.2000;31.12.2001;31.12.2002;31.12.2003;31.12.2004;31.12.2005;31.12.2006;31.12.2007 Baden-Württemberg;35752,50;35751,76;35751,85;35751,63;35751,36;35751,36;35751,64;35751,64;35751,65;35751,64;35751,65;35751,50;35751,40 Bayern;70550,87;70550,87;70548,00;70547,96;70547,81;70547,85;70549,93;70549,32;70549,19;70549,44;70551,57;70551,57;70551,56 Berlin;890,82;890,85;890,77;890,22;891,41;891,69;891,76;891,75;891,75;891,82;891,85;891,02;891,02 Thüringen;16171,12;16170,88;16171,57;16171,70;16171,85;16171,98;16171,94;16172,21;16172,14;16172,08;16172,10;16172,14;16172,10 Insgesamt;357022,31;357021,43;357020,79;357022,17;357020,22;357021,54;357022,90;357026,55;357030,32;357045,64;357092,90;357114,25;357104,07 _____ Gebietsfläche: "Berlin (1995-2000): bezogen auf den Gebietsstand 01.01.2001;" Rheinland-Pfalz: Landessumme ab 2004 einschl. des gemein- " schaftlichen deutsch-luxemburgischen Hoheitsgebiets;" Sachsen (1995): bezogen auf den Gebietsstand 01.01.1996. _____ (C)opyright Statistisches Bundesamt, Wiesbaden 2010 Stand: 22.04.2010 / 21:32:50""" GEBURTENZIFFERN_12612_0102 = """GENESIS-Tabelle: Temporär Geburtenziffern: Deutschland, Jahre, Altersjahre;;;;;;;; Statistik der Geburten;;;;;;;; Deutschland;;;;;;;; Lebendgeborene je 1000 Frauen (Anzahl);;;;;;;; ;Jahr;Jahr;Jahr;Jahr;Jahr;Jahr;Jahr;Jahr ;2001;2002;2003;2004;2005;2006;2007;2008 15 Jahre;1,0;1,0;0,9;0,8;0,8;0,8;0,8;0,8 16 Jahre;3,1;3,2;3,0;2,8;2,6;2,6;2,5;2,6 17 Jahre;7,5;7,7;7,1;6,4;6,2;5,8;5,5;5,7 47 Jahre;0,2;0,1;0,2;0,2;0,2;0,2;0,2;0,2 48 Jahre;0,0;0,1;0,1;0,1;0,1;0,1;0,1;0,1 49 Jahre;0,0;0,0;0,0;0,0;0,1;0,1;0,1;0,1 _____ Durchschnittliches Alter: Differenz zwischen Geburtsjahr des Kindes und Geburtsjahr der Mutter. _____ (C)opyright Statistisches Bundesamt, Wiesbaden 2010 Stand: 10.04.2010 / 21:25:57""" class tableParserTest(unittest.TestCase): def _make_parser(self, data, table_id): structure, variables = table.load_structure(table_id) return table.tableParser(structure, variables, data) def setUp(self): pass def test_parse_geburtenziffern(self): parser = self._make_parser(GEBURTENZIFFERN_12612_0102, "12612-0102") parser.parse() assert False if __name__ == '__main__': unittest.main()
mit
-8,605,079,979,203,655,000
38.529412
143
0.709341
false
hexpl0it/plugin.video.genesi-ita
resources/lib/libraries/js2py/prototypes/jsarray.py
1
14657
def to_arr(this): """Returns Python array from Js array""" return [this.get(str(e)) for e in xrange(len(this))] ARR_STACK = set({}) class ArrayPrototype: def toString(): # this function is wrong but I will leave it here fore debugging purposes. func = this.get('join') if not func.is_callable(): @this.Js def func(): return '[object %s]'%this.Class return func.call(this, ()) def toLocaleString(): array = this.to_object() arr_len = array.get('length').to_uint32() # separator is simply a comma ',' if not arr_len: return '' res = [] for i in xrange(arr_len): element = array[str(i)] if element.is_undefined() or element.is_null(): res.append('') else: cand = element.to_object() str_func = element.get('toLocaleString') if not str_func.is_callable(): raise this.MakeError('TypeError', 'toLocaleString method of item at index %d is not callable'%i) res.append(element.callprop('toLocaleString').value) return ','.join(res) def concat(): array = this.to_object() A = this.Js([]) items = [array] items.extend(to_arr(arguments)) n = 0 for E in items: if E.Class=='Array': k = 0 e_len = len(E) while k<e_len: if E.has_property(str(k)): A.put(str(n), E.get(str(k))) n+=1 k+=1 else: A.put(str(n), E) n+=1 return A def join(separator): ARR_STACK.add(this) array = this.to_object() arr_len = array.get('length').to_uint32() separator = ',' if separator.is_undefined() else separator.to_string().value elems = [] for e in xrange(arr_len): elem = array.get(str(e)) if elem in ARR_STACK: s = '' else: s = elem.to_string().value elems.append(s if not (elem.is_undefined() or elem.is_null()) else '') res = separator.join(elems) ARR_STACK.remove(this) return res def pop(): #todo check array = this.to_object() arr_len = array.get('length').to_uint32() if not arr_len: array.put('length', this.Js(arr_len)) return None ind = str(arr_len-1) element = array.get(ind) array.delete(ind) array.put('length', this.Js(arr_len-1)) return element def push(item): # todo check array = this.to_object() arr_len = array.get('length').to_uint32() to_put = arguments.to_list() i = arr_len for i, e in enumerate(to_put, arr_len): array.put(str(i), e) if to_put: i+=1 array.put('length', this.Js(i)) return i def reverse(): array = this.to_object() # my own algorithm vals = to_arr(array) has_props = [array.has_property(str(e)) for e in xrange(len(array))] vals.reverse() has_props.reverse() for i, val in enumerate(vals): if has_props[i]: array.put(str(i), val) else: array.delete(str(i)) return array def shift(): #todo check array = this.to_object() arr_len = array.get('length').to_uint32() if not arr_len: array.put('length', this.Js(0)) return None first = array.get('0') for k in xrange(1, arr_len): from_s, to_s = str(k), str(k-1) if array.has_property(from_s): array.put(to_s, array.get(from_s)) else: array.delete(to) array.delete(str(arr_len-1)) array.put('length', this.Js(str(arr_len-1))) return first def slice(start, end): # todo check array = this.to_object() arr_len = array.get('length').to_uint32() relative_start = start.to_int() k = max((arr_len + relative_start), 0) if relative_start<0 else min(relative_start, arr_len) relative_end = arr_len if end.is_undefined() else end.to_int() final = max((arr_len + relative_end), 0) if relative_end<0 else min(relative_end, arr_len) res = [] n = 0 while k<final: pk = str(k) if array.has_property(pk): res.append(array.get(pk)) k += 1 n += 1 return res def sort(cmpfn): if not this.Class in {'Array', 'Arguments'}: return this.to_object() # do nothing arr = [this.get(str(i)) for i in xrange(len(this))] if not arr: return this if not cmpfn.is_callable(): cmpfn = None cmp = lambda a,b: sort_compare(a, b, cmpfn) arr.sort(cmp=cmp) for i in xrange(len(arr)): this.put(unicode(i), arr[i]) return this def splice(start, deleteCount): # 1-8 array = this.to_object() arr_len = array.get('length').to_uint32() relative_start = start.to_int() actual_start = max((arr_len + relative_start),0) if relative_start<0 else min(relative_start, arr_len) actual_delete_count = min(max(deleteCount.to_int(),0 ), arr_len - actual_start) k = 0 A = this.Js([]) # 9 while k<actual_delete_count: if array.has_property(str(actual_start+k)): A.put(str(k), array.get(str(actual_start+k))) k += 1 # 10-11 items = to_arr(arguments)[2:] items_len = len(items) # 12 if items_len<actual_delete_count: k = actual_start while k < (arr_len-actual_delete_count): fr = str(k+actual_delete_count) to = str(k+items_len) if array.has_property(fr): array.put(to, array.get(fr)) else: array.delete(to) k += 1 k = arr_len while k > (arr_len - actual_delete_count + items_len): array.delete(str(k-1)) k -= 1 # 13 elif items_len>actual_delete_count: k = arr_len - actual_delete_count while k>actual_start: fr = str(k + actual_delete_count - 1) to = str(k + items_len - 1) if array.has_property(fr): array.put(to, array.get(fr)) else: array.delete(to) k -= 1 # 14-17 k = actual_start while items: E = items.pop(0) array.put(str(k), E) k += 1 array.put('length', this.Js(arr_len - actual_delete_count + items_len)) return A def unshift(): array = this.to_object() arr_len = array.get('length').to_uint32() argCount = len(arguments) k = arr_len while k > 0: fr = str(k - 1) to = str(k + argCount - 1) if array.has_property(fr): array.put(to, array.get(fr)) else: array.delete(to) k -= 1 j = 0 items = to_arr(arguments) while items: E = items.pop(0) array.put(str(j), E) j += 1 array.put('length', this.Js(arr_len + argCount)) return arr_len + argCount def indexOf(searchElement): array = this.to_object() arr_len = array.get('length').to_uint32() if arr_len == 0: return -1 if len(arguments)>1: n = arguments[1].to_int() else: n = 0 if n >= arr_len: return -1 if n >= 0: k = n else: k = arr_len - abs(n) if k < 0: k = 0 while k < arr_len: if array.has_property(str(k)): elementK = array.get(str(k)) if searchElement.strict_equality_comparison(elementK): return k k += 1 return -1 def lastIndexOf(searchElement): array = this.to_object() arr_len = array.get('length').to_uint32() if arr_len == 0: return -1 if len(arguments)>1: n = arguments[1].to_int() else: n = arr_len - 1 if n >= 0: k = min(n, arr_len-1) else: k = arr_len - abs(n) while k >= 0: if array.has_property(str(k)): elementK = array.get(str(k)) if searchElement.strict_equality_comparison(elementK): return k k -= 1 return -1 def every(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') T = arguments[1] k = 0 while k<arr_len: if array.has_property(str(k)): kValue = array.get(str(k)) if not callbackfn.call(T, (kValue, this.Js(k), array)).to_boolean().value: return False k += 1 return True def some(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') T = arguments[1] k = 0 while k<arr_len: if array.has_property(str(k)): kValue = array.get(str(k)) if callbackfn.call(T, (kValue, this.Js(k), array)).to_boolean().value: return True k += 1 return False def forEach(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') T = arguments[1] k = 0 while k<arr_len: if array.has_property(str(k)): kValue = array.get(str(k)) callbackfn.call(T, (kValue, this.Js(k), array)) k+=1 def map(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') T = arguments[1] A = this.Js([]) k = 0 while k<arr_len: Pk = str(k) if array.has_property(Pk): kValue = array.get(Pk) mappedValue = callbackfn.call(T, (kValue, this.Js(k), array)) A.define_own_property(Pk, {'value': mappedValue, 'writable': True, 'enumerable': True, 'configurable': True}) k += 1 return A def filter(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') T = arguments[1] res = [] k = 0 while k<arr_len: if array.has_property(str(k)): kValue = array.get(str(k)) if callbackfn.call(T, (kValue, this.Js(k), array)).to_boolean().value: res.append(kValue) k += 1 return res # converted to js array automatically def reduce(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') if not arr_len and len(arguments)<2: raise this.MakeError('TypeError', 'Reduce of empty array with no initial value') k = 0 if len(arguments)>1: # initial value present accumulator = arguments[1] else: kPresent = False while not kPresent and k<arr_len: kPresent = array.has_property(str(k)) if kPresent: accumulator = array.get(str(k)) k += 1 if not kPresent: raise this.MakeError('TypeError', 'Reduce of empty array with no initial value') while k<arr_len: if array.has_property(str(k)): kValue = array.get(str(k)) accumulator = callbackfn.call(this.undefined, (accumulator, kValue, this.Js(k), array)) k += 1 return accumulator def reduceRight(callbackfn): array = this.to_object() arr_len = array.get('length').to_uint32() if not callbackfn.is_callable(): raise this.MakeError('TypeError', 'callbackfn must be a function') if not arr_len and len(arguments)<2: raise this.MakeError('TypeError', 'Reduce of empty array with no initial value') k = arr_len - 1 if len(arguments)>1: # initial value present accumulator = arguments[1] else: kPresent = False while not kPresent and k>=0: kPresent = array.has_property(str(k)) if kPresent: accumulator = array.get(str(k)) k -= 1 if not kPresent: raise this.MakeError('TypeError', 'Reduce of empty array with no initial value') while k>=0: if array.has_property(str(k)): kValue = array.get(str(k)) accumulator = callbackfn.call(this.undefined, (accumulator, kValue, this.Js(k), array)) k -= 1 return accumulator def sort_compare(a, b, comp): if a is None: if b is None: return 0 return 1 if b is None: if a is None: return 0 return -1 if a.is_undefined(): if b.is_undefined(): return 0 return 1 if b.is_undefined(): if a.is_undefined(): return 0 return -1 if comp is not None: res = comp.call(a.undefined, (a, b)) return res.to_int() x, y = a.to_string(), b.to_string() if x<y: return -1 elif x>y: return 1 return 0
gpl-3.0
2,274,216,971,756,242,200
32.011261
116
0.492324
false
Battleroid/yanker
yanker/yanker.py
1
2508
""" Yanker Usage: yanker [--threads=<tnum>] """ __version__ = '1.0.1' import Queue import threading import youtube_dl as ydl import pyperclip as clip import time from docopt import docopt class ErrLogger(object): def debug(self, msg): pass def warning(self, msg): pass def error(self, msg): print msg class Worker(threading.Thread): def __init__(self, tasks): threading.Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: vid = self.tasks.get() vid.download() self.tasks.task_done() class Video: def progress(self, s): if s['status'] == 'finished': print 'Finished {}'.format(s['filename']) def __init__(self, url, opts={}): self.url = url self.ydl_opts = { 'progress_hooks': [self.progress], 'logger': ErrLogger() } self.ydl_opts.update(opts) def download(self): print 'Downloading: {}'.format(self.url) with ydl.YoutubeDL(self.ydl_opts) as y: try: y.download([self.url]) except ydl.DownloadError: print 'Unsupported URL, skipping' class Watcher: def __init__(self, urls=[], threads=2): self.queue = Queue.Queue(0) self.threads = threads self.stopped = False self.grabbed_urls = set([]) for _ in range(threads): Worker(self.queue) def run(self): recent = '' while not self.stopped: current = clip.paste() if recent != current: recent = current if current.startswith(('http://', 'https://',)) and current not in self.grabbed_urls: print 'Added: {}'.format(current) self.grabbed_urls.add(current) self.queue.put(Video(current)) elif current in self.grabbed_urls: print 'Already grabbed {}'.format(current) time.sleep(0.25) def run(): args = docopt(__doc__, version='Yanker {}'.format(__version__)) threads = args['--threads'] if not threads: threads = 2 else: threads = int(threads) print 'Starting Yanker with {} threads...'.format(threads) watch = Watcher(threads=threads) try: watch.run() except KeyboardInterrupt: print 'Stopping...' watch.stopped = True
mit
-6,076,938,484,919,029,000
24.08
101
0.539474
false
Traumflug/Teacup_Firmware
configtool/calcscrew.py
1
9429
import wx from configtool.data import BSIZESMALL, reFloat, offsetChLabel, offsetTcLabel class CalcScrew(wx.Dialog): def __init__(self, parent, font, cbUse): wx.Dialog.__init__( self, parent, wx.ID_ANY, "Steps calculator for screw driven axes", size=(400, 204), ) self.SetFont(font) self.Bind(wx.EVT_CLOSE, self.onExit) self.use = cbUse labelWidth = 150 hsz = wx.BoxSizer(wx.HORIZONTAL) hsz.Add((10, 10)) sz = wx.BoxSizer(wx.VERTICAL) sz.Add((10, 10)) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Step Angle:", size=(labelWidth, -1), style=wx.ALIGN_RIGHT ) st.SetFont(font) lsz.Add(st, 1, wx.TOP, offsetChLabel) lsz.Add((5, 5)) stepAngles = [ "1.8 (200 per revolution)", "0.9 (400 per revolution)", "7.5 (48 per revolution)", ] self.stepAngleValues = [200, 400, 48] tc = wx.Choice(self, wx.ID_ANY, choices=stepAngles) tc.SetFont(font) tc.SetSelection(0) tc.Bind(wx.EVT_CHOICE, self.onChoice) lsz.Add(tc) tc.SetToolTip("Step angle. Depends on your type of stepper motor.") self.tcStep = tc sz.Add(lsz) sz.Add((10, 10)) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Microstepping:", size=(labelWidth, -1), style=wx.ALIGN_RIGHT, ) st.SetFont(font) lsz.Add(st, 1, wx.TOP, offsetChLabel) lsz.Add((5, 5)) microStepping = [ "1 - full step", "1/2 - half step", "1/4 - quarter step", "1/8", "1/16", "1/32", "1/64", "1/128", ] self.microSteppingValues = [1, 2, 4, 8, 16, 32, 64, 128] tc = wx.Choice(self, wx.ID_ANY, choices=microStepping) tc.SetFont(font) tc.Bind(wx.EVT_CHOICE, self.onChoice) tc.SetSelection(4) lsz.Add(tc) tc.SetToolTip( "Microstepping. Most boards allow to change this by " "setting jumpers. The value here must match the " "setting on the board in conjunction with the type " "of stepper driver chip." ) self.tcMicroStepping = tc sz.Add(lsz) sz.Add((10, 10)) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Screw Pitch (mm/rev):", size=(labelWidth, -1), style=wx.ALIGN_RIGHT, ) st.SetFont(font) lsz.Add(st, 1, wx.TOP, offsetTcLabel) lsz.Add((5, 5)) tc = wx.TextCtrl(self, wx.ID_ANY, "2", style=wx.TE_RIGHT) tc.SetFont(font) tc.Bind(wx.EVT_TEXT, self.onTextCtrlFloat) lsz.Add(tc) tc.SetToolTip("Screw pitch. Defined by the pitch of the screw.") self.tcScrewPitch = tc lsz.Add((5, 5)) screwPresets = [ "-", "M8 - metric (1.25 mm/rev)", "M6 - metric (1 mm/rev)", "M5 - metric (0.8 mm/rev)", "12 (12 mm/rev)", "16 (16 mm/rev)", "25 (25 mm/rev)", '5/15"-18 imperial coarse (1.41111 mm/rev)', '3/16"-20 imperial (1.270 mm/rev)', '1/4"-16 ACME (1.5875 mm/rev)', ] self.screwPresetValues = [ -1, 1.25, 1.00, 0.8, 12.0, 16.0, 25.0, 1.41111, 1.270, 1.5875, ] tc = wx.Choice(self, wx.ID_ANY, choices=screwPresets) tc.SetFont(font) tc.SetSelection(0) tc.Bind(wx.EVT_CHOICE, self.onPresetChoice) lsz.Add(tc) tc.SetToolTip("Screw pitch presets.") self.tcPresets = tc sz.Add(lsz) sz.Add((10, 10)) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Gear Ratio:", size=(labelWidth, -1), style=wx.ALIGN_RIGHT ) st.SetFont(font) lsz.Add(st, 1, wx.TOP, offsetTcLabel) lsz.Add((5, 5)) tc = wx.TextCtrl(self, wx.ID_ANY, "1", size=(40, -1), style=wx.TE_RIGHT) tc.SetFont(font) tc.Bind(wx.EVT_TEXT, self.onTextCtrlFloat) lsz.Add(tc) tc.SetToolTip("Gear ratio. 1:1 if there is no gear.") self.tcRatioTop = tc lsz.Add((5, 5)) st = wx.StaticText(self, wx.ID_ANY, ":") st.SetFont(font) lsz.Add(st) lsz.Add((5, 5)) tc = wx.TextCtrl(self, wx.ID_ANY, "1", size=(40, -1), style=wx.TE_RIGHT) tc.SetFont(font) tc.Bind(wx.EVT_TEXT, self.onTextCtrlFloat) lsz.Add(tc) tc.SetToolTip("Gear ratio. 1:1 if there is no gear.") self.tcRatioBottom = tc sz.Add(lsz) sz.Add((30, 30)) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Result:", size=(labelWidth, -1), style=wx.ALIGN_RIGHT ) st.SetFont(font) lsz.Add(st) lsz.Add((5, 5)) tc = wx.StaticText(self, wx.ID_ANY, "", size=(300, -1), style=wx.ALIGN_LEFT) tc.SetFont(font) lsz.Add(tc) self.tcResult = tc sz.Add(lsz) lsz = wx.BoxSizer(wx.HORIZONTAL) st = wx.StaticText( self, wx.ID_ANY, "Resolution:", size=(labelWidth, -1), style=wx.ALIGN_RIGHT ) st.SetFont(font) lsz.Add(st) lsz.Add((5, 5)) tc = wx.StaticText(self, wx.ID_ANY, "", size=(300, -1), style=wx.ALIGN_LEFT) tc.SetFont(font) lsz.Add(tc) self.tcResolution = tc sz.Add(lsz) sz.Add((20, 20)) bsz = wx.BoxSizer(wx.HORIZONTAL) b = wx.Button(self, wx.ID_ANY, "Use for X", size=BSIZESMALL) b.SetFont(font) self.Bind(wx.EVT_BUTTON, self.onUseForX, b) bsz.Add(b) self.bUseForX = b bsz.Add((5, 5)) b = wx.Button(self, wx.ID_ANY, "Use for Y", size=BSIZESMALL) b.SetFont(font) self.Bind(wx.EVT_BUTTON, self.onUseForY, b) bsz.Add(b) self.bUseForY = b bsz.Add((5, 5)) b = wx.Button(self, wx.ID_ANY, "Use for Z", size=BSIZESMALL) b.SetFont(font) self.Bind(wx.EVT_BUTTON, self.onUseForZ, b) bsz.Add(b) self.bUseForZ = b bsz.Add((5, 5)) b = wx.Button(self, wx.ID_ANY, "Use for E", size=BSIZESMALL) b.SetFont(font) self.Bind(wx.EVT_BUTTON, self.onUseForE, b) bsz.Add(b) self.bUseForE = b sz.Add(bsz, flag=wx.ALIGN_CENTER_HORIZONTAL) sz.Add((10, 10)) hsz.Add(sz) hsz.Add((10, 10)) self.enableUseButtons(False) self.SetSizer(hsz) self.Fit() self.calculate() def calculate(self): self.tcResult.SetLabel("") self.tcResolution.SetLabel("") self.enableUseButtons(False) s = self.tcStep.GetSelection() sv = self.stepAngleValues[s] try: sp = float(self.tcScrewPitch.GetValue()) except: return try: ratioA = float(self.tcRatioTop.GetValue()) except: return try: ratioB = float(self.tcRatioBottom.GetValue()) except: return s = self.tcMicroStepping.GetSelection() msv = self.microSteppingValues[s] ratio = ratioA / ratioB steps = sv * msv resultmm = steps / sp / ratio self.result = int(resultmm * 1000.0) self.tcResult.SetLabel("%d steps/m (%.3f steps/mm)" % (self.result, resultmm)) self.tcResolution.SetLabel("%.3f micrometers" % (1.0 / resultmm * 1000.0)) self.enableUseButtons(True) def enableUseButtons(self, flag): self.bUseForX.Enable(flag) self.bUseForY.Enable(flag) self.bUseForZ.Enable(flag) self.bUseForE.Enable(flag) def onUseForX(self, evt): self.use("STEPS_PER_M_X", self.result) def onUseForY(self, evt): self.use("STEPS_PER_M_Y", self.result) def onUseForZ(self, evt): self.use("STEPS_PER_M_Z", self.result) def onUseForE(self, evt): self.use("STEPS_PER_M_E", self.result) def onPresetChoice(self, evt): s = self.tcPresets.GetSelection() sv = self.screwPresetValues[s] if sv < 0: return s = "%f" % sv s = s.rstrip("0") if s[-1] == ".": s += "0" self.tcScrewPitch.SetValue(s) def onChoice(self, evt): self.calculate() def onTextCtrlFloat(self, evt): tc = evt.GetEventObject() w = tc.GetValue().strip() if w == "": valid = False else: m = reFloat.match(w) if m: valid = True else: valid = False if valid: tc.SetBackgroundColour(wx.SystemSettings.GetColour(wx.SYS_COLOUR_WINDOW)) else: tc.SetBackgroundColour("pink") tc.Refresh() self.calculate() evt.Skip() def onExit(self, evt): self.EndModal(wx.ID_OK)
gpl-2.0
-6,171,746,121,561,787,000
26.814159
88
0.509492
false
BadWizard/Inflation
SPF/source/get_clean_GDP.py
1
3646
def get_clean_HICP(data_dir="../data/", output_dir="../clean_data/", year=2016, quarter=1): ''' This function takes a raw csv file from SPF and creates a new csv file with forecasts only for HICP ''' import pandas as pd import numpy as np import os # this is fixed list, will use it to create a new target months = ['Dec', 'Mar', 'Jun', 'Sep'] # this is the raw file name raw_file = str(year) + 'Q' + str(quarter) + '.csv' # this is the name of the file that will be saved output_file = 'HICP' + raw_file # full path of the input file fname = data_dir + raw_file # check if the input file exists if not os.path.isfile(fname): print('File ' + fname + 'is not available') else: raw_df = pd.read_csv(fname, header=1) # find the row where the growth expectations start dum = raw_df[raw_df['TARGET_PERIOD'] == 'GROWTH EXPECTATIONS; YEAR-ON-YEAR CHANGE IN REAL GDP'].index[0] mask_columns = ~raw_df.columns.str.contains('Unnamed') df = raw_df.iloc[0:dum-1, mask_columns] df = df.rename(columns={'TARGET_PERIOD':'target','FCT_SOURCE':'id','POINT':'point', 'TN1_0':'[-2.0,-1.1]','FN1_0TN0_6':'[-1.0,-0.6]', 'FN0_5TN0_1':'[-0.5,-0.1]','F0_0T0_4':'[0.0,0.4]', 'F0_5T0_9':'[0.5,0.9]','F1_0T1_4':'[1.0,1.4]', 'F1_5T1_9':'[1.5,1.9]','F2_0T2_4':'[2.0,2.4]', 'F2_5T2_9':'[2.5,2.9]','F3_0T3_4':'[3.0,3.4]', 'F3_5T3_9':'[3.5,3.9]','F4_0':'[4.0,5.0]'}) # remove rows where point is missing maskNaN = df.point.isnull() df = df[~maskNaN] df.fillna(0, inplace=True) for colname in df.columns[3:]: df[colname] = df[colname].astype('float') # create a new target column mask_t0 = str(year) mask_t1 = str(year+1) mask_t2 = str(year+2) if quarter < 3: mask_t4or5 = str(year+4) else: mask_t4or5 = str(year+5) if quarter == 1: mask_Rt1 = str(year) + months[quarter-1] mask_Rt2 = str(year+1) + months[quarter-1] else: mask_Rt1 = str(year+1) + months[quarter-1] mask_Rt2 = str(year+2) + months[quarter-1] df.loc[df.loc[:, 'target'] == mask_t0, 'targetNew'] = 't' df.loc[df.loc[:, 'target'] == mask_t1, 'targetNew'] = 't+1' df.loc[df.loc[:, 'target'] == mask_t2, 'targetNew'] = 't+2' df.loc[df.loc[:, 'target'] == mask_t4or5, 'targetNew'] = 't+4' df.loc[df.loc[:, 'target'] == mask_Rt1, 'targetNew'] = 'roll 1' df.loc[df.loc[:, 'target'] == mask_Rt2, 'targetNew'] = 'roll 2' df['source'] = str(year) + '-Q' + str(quarter) df = df[['source', 'target', 'targetNew', 'id', 'point', '[-2.0,-1.1]', '[-1.0,-0.6]', '[-0.5,-0.1]', '[0.0,0.4]', '[0.5,0.9]', '[1.0,1.4]', '[1.5,1.9]', '[2.0,2.4]', '[2.5,2.9]', '[3.0,3.4]', '[3.5,3.9]', '[4.0,5.0]']] # save to a csv file df.to_csv(output_dir + output_file, index=False) def main(): data_dir = "../data/" year = 2016 quarter = 1 output_dir = "../clean_data/" for y in [2015, 2016]: for q in [1, 2, 3, 4]: get_clean_HICP(data_dir=data_dir, output_dir=output_dir, year=y, quarter=q) main()
mit
8,939,961,401,442,536,000
36.204082
91
0.472024
false
factorlibre/carrier-delivery
delivery_carrier_mrw/model/stock.py
1
8462
# -*- encoding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2015 FactorLibre (http://www.factorlibre.com) # Hugo Santos <[email protected]> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import urllib from datetime import datetime from openerp import models, fields, api, exceptions from openerp.tools.translate import _ from ..webservice.mrw_api import MrwEnvio class StockPicking(models.Model): _inherit = 'stock.picking' @api.model def _get_mrw_service_type(self): return [ ('0000', 'Urgente 10'), ('0005', 'Urgente Hoy'), ('0100', 'Urgente 12'), ('0110', 'Urgente 14'), ('0120', 'Urgente 22'), ('0200', 'Urgente 19'), ('0205', 'Urgente 19 Expedicion'), ('0210', 'Urgente 19 Mas 40 kilos'), ('0220', 'Urgente 19 Portugal'), ('0230', 'Bag 19'), ('0235', 'Bag 14'), ('0300', 'Economico'), ('0310', 'Economico Mas 40 Kilos'), ('0350', 'Economico Interinsular'), ('0400', 'Express Documentos'), ('0450', 'Express 2 Kilos'), ('0480', 'Caja Express 3 Kilos'), ('0490', 'Documentos 14'), ('0800', 'Ecommerce') ] mrw_service_type = fields.Selection( '_get_mrw_service_type', string='Mrw Service') mrw_frequence = fields.Selection( (('1', 'Frecuencia 1'), ('2', 'Frecuencia 2')), string='Mrw Frequence') @api.multi def _mrw_transm_envio_request(self, mrw_api): self.ensure_one() client = mrw_api.client transm_envio = client.factory.create('TransmEnvioRequest') warehouse_address = self.picking_type_id.warehouse_id.partner_id pickup_address = transm_envio.DatosRecogida.Direccion pickup_address.Via = warehouse_address.street if warehouse_address.street2: pickup_address.Resto = warehouse_address.street2 # TODO: Comprobar que hacer con el resto de codigos postales # Llevan un formato especial por país # - España: Poner los 5 dígitos. (Ej: 05200 para Ávila) # - Portugal: Poner sólo los 4 primeros dígitos de los 7. # (Ej: 1200 para Lisboa) # - Andorra: Deben ser 5 dígitos, por lo que se pondrá un 0 delante del # mismo (p. ej 00500 para Andorra la Vella) pickup_address.CodigoPostal = warehouse_address.zip.zfill(5) pickup_address.Poblacion = warehouse_address.city pickup_address.Provincia = warehouse_address.state_id.name or '' pickup_address.CodigoPais = warehouse_address.country_id.code or '' transm_envio.DatosRecogida.Nif = warehouse_address.vat or '' transm_envio.DatosRecogida.Nombre = warehouse_address.name transm_envio.DatosRecogida.Telefono = warehouse_address.phone or '' shipping_address = transm_envio.DatosEntrega.Direccion shipping_address.Via = self.partner_id.street shipping_address.Resto = self.partner_id.street2 or '' shipping_address.CodigoPostal = self.partner_id.zip shipping_address.Poblacion = self.partner_id.city shipping_address.Provincia = self.partner_id.state_id.name or '' shipping_address.CodigoPais = self.partner_id.country_id.name or '' transm_envio.DatosEntrega.Nif = self.partner_id.vat or '' transm_envio.DatosEntrega.Nombre = self.partner_id.name or '' transm_envio.DatosEntrega.Telefono = self.partner_id.phone or '' # Datos Servicio service_data = transm_envio.DatosServicio service_data.Fecha = datetime.strftime( fields.Datetime.from_string(self.date_done), '%d/%m/%Y') service_data.Referencia = self.name service_data.EnFranquicia = 'N' service_data.CodigoServicio = self.mrw_service_type service_data.NumeroBultos = self.number_of_packages or 1 service_data.Peso = self.weight or 1 if self.mrw_frequence: service_data.Frecuencia = self.mrw_frequence # TODO: Servicio Rembolso # Reembolso: indicador opcional de reembolso. Valores posibles: # - N: (default) Sin reembolso. # - O: Con reembolso comisión en origen. # - D: Con reembolso comisión en destino. # ImporteReembolso: # - importe nominal del reembolso (para envíos con reembolso) if self.partner_id.email: notification_request = client.factory.create('NotificacionRequest') notification_request.CanalNotificacion = "1" notification_request.TipoNotificacion = "4" notification_request.MailSMS = self.partner_id.email service_data.Notificaciones.NotificacionRequest.append( notification_request) return transm_envio @api.multi def _mrw_etiqueta_envio_request(self, mrw_api, shipping_number): self.ensure_one() client = mrw_api.client label_factory = client.factory.create('EtiquetaEnvioRequest') label_factory.NumeroEnvio = shipping_number label_factory.ReportTopMargin = "1100" label_factory.ReportLeftMargin = "650" return label_factory @api.multi def _generate_mrw_label(self, package_ids=None): self.ensure_one() if not self.carrier_id.mrw_config_id: raise exceptions.Warning(_('No MRW Config defined in carrier')) if not self.picking_type_id.warehouse_id.partner_id: raise exceptions.Warning( _('Please define an address in the %s warehouse') % ( self.warehouse_id.name)) mrw_api = MrwEnvio(self.carrier_id.mrw_config_id) client = mrw_api.client transm_envio = self._mrw_transm_envio_request(mrw_api) response = client.service.TransmEnvio(transm_envio) if response.Estado != '1' and not response.NumeroEnvio: raise exceptions.Warning(response.Mensaje) label_factory = self._mrw_etiqueta_envio_request(mrw_api, response.NumeroEnvio) label_response = client.service.EtiquetaEnvio(label_factory) if label_response.Estado != '1': raise exceptions.Warning(response.Mensaje) label = { 'file': label_response.EtiquetaFile.decode('base64'), 'file_type': 'pdf', 'name': response.NumeroEnvio + '.pdf', } return [label] @api.multi def _get_mrw_label_from_url(self, shipping_number): self.ensure_one() mrw_config = self.carrier_id.mrw_config_id url = "http://sagec.mrw.es" if mrw_config.is_test: url = "http://sagec-test.mrw.es" params = { 'Franq': mrw_config.franchise_code, 'Ab': mrw_config.subscriber_code, 'Dep': mrw_config.department_code or '', 'Usr': mrw_config.username, 'Pwd': mrw_config.password, 'NumEnv': shipping_number } url_params = urllib.urlencode(params) # Generar etiqueta en Panel panel_url = u"{0}/Panel.aspx?{1}".format(url, url_params) return panel_url @api.multi def generate_shipping_labels(self, package_ids=None): """ Add label generation for MRW """ self.ensure_one() if self.carrier_id.type == 'mrw': return self._generate_mrw_label(package_ids=package_ids) return super(StockPicking, self).generate_shipping_labels( package_ids=package_ids)
agpl-3.0
-7,478,398,197,497,246,000
40.024272
79
0.609869
false
xlevus/python-diana
docs/source/conf.py
1
9861
# -*- coding: utf-8 -*- # # Diana documentation build configuration file, created by # sphinx-quickstart on Wed Jun 22 12:48:46 2016. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # # import os # import sys # sys.path.insert(0, os.path.abspath('.')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. # # needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ "sphinx.ext.autodoc", "sphinx.ext.doctest", "sphinx.ext.intersphinx", "sphinx.ext.todo", "sphinx.ext.coverage", "sphinx.ext.viewcode", "sphinx.ext.githubpages", ] # Add any paths that contain templates here, relative to this directory. templates_path = ["_templates"] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # # source_suffix = ['.rst', '.md'] source_suffix = ".rst" # The encoding of source files. # # source_encoding = 'utf-8-sig' # The master toctree document. master_doc = "index" # General information about the project. project = u"Diana" copyright = u"2016, Chris Targett" author = u"Chris Targett" # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = u"0.1.0" # The full version, including alpha/beta/rc tags. release = u"0.1.0" # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: # # today = '' # # Else, today_fmt is used as the format for a strftime call. # # today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This patterns also effect to html_static_path and html_extra_path exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all # documents. # # default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. # # add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). # # add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. # # show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = "sphinx" # A list of ignored prefixes for module index sorting. # modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. # keep_warnings = False # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = True # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = "alabaster" # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # # html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. # html_theme_path = [] # The name for this set of Sphinx documents. # "<project> v<release> documentation" by default. # # html_title = u'Diana v0.1.0' # A shorter title for the navigation bar. Default is the same as html_title. # # html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. # # html_logo = None # The name of an image file (relative to this directory) to use as a favicon of # the docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. # # html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ["_static"] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. # # html_extra_path = [] # If not None, a 'Last updated on:' timestamp is inserted at every page # bottom, using the given strftime format. # The empty string is equivalent to '%b %d, %Y'. # # html_last_updated_fmt = None # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. # # html_use_smartypants = True # Custom sidebar templates, maps document names to template names. # # html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. # # html_additional_pages = {} # If false, no module index is generated. # # html_domain_indices = True # If false, no index is generated. # # html_use_index = True # If true, the index is split into individual pages for each letter. # # html_split_index = False # If true, links to the reST sources are added to the pages. # # html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. # # html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. # # html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. # # html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). # html_file_suffix = None # Language to be used for generating the HTML full-text search index. # Sphinx supports the following languages: # 'da', 'de', 'en', 'es', 'fi', 'fr', 'hu', 'it', 'ja' # 'nl', 'no', 'pt', 'ro', 'ru', 'sv', 'tr', 'zh' # # html_search_language = 'en' # A dictionary with options for the search language support, empty by default. # 'ja' uses this config value. # 'zh' user can custom change `jieba` dictionary path. # # html_search_options = {'type': 'default'} # The name of a javascript file (relative to the configuration directory) that # implements a search results scorer. If empty, the default will be used. # # html_search_scorer = 'scorer.js' # Output file base name for HTML help builder. htmlhelp_basename = "Dianadoc" # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). # # 'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). # # 'pointsize': '10pt', # Additional stuff for the LaTeX preamble. # # 'preamble': '', # Latex figure (float) alignment # # 'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, "Diana.tex", u"Diana Documentation", u"Chris Targett", "manual"), ] # The name of an image file (relative to this directory) to place at the top of # the title page. # # latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. # # latex_use_parts = False # If true, show page references after internal links. # # latex_show_pagerefs = False # If true, show URL addresses after external links. # # latex_show_urls = False # Documents to append as an appendix to all manuals. # # latex_appendices = [] # If false, no module index is generated. # # latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [(master_doc, "diana", u"Diana Documentation", [author], 1)] # If true, show URL addresses after external links. # # man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ( master_doc, "Diana", u"Diana Documentation", author, "Diana", "One line description of project.", "Miscellaneous", ), ] # Documents to append as an appendix to all manuals. # # texinfo_appendices = [] # If false, no module index is generated. # # texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. # # texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. # # texinfo_no_detailmenu = False # Example configuration for intersphinx: refer to the Python standard library. intersphinx_mapping = {"https://docs.python.org/": None}
mit
-4,637,041,529,167,908,000
27.749271
82
0.689585
false
alphagov/digitalmarketplace-admin-frontend
tests/app/main/views/test_outcomes.py
1
9064
import csv import mock import pytest from dmtestutils.api_model_stubs import FrameworkStub from ...helpers import LoggedInApplicationTest class TestDirectAwardView(LoggedInApplicationTest): def setup_method(self, method): super().setup_method(method) self.data_api_client_patch = mock.patch('app.main.views.outcomes.data_api_client', autospec=True) self.data_api_client = self.data_api_client_patch.start() def teardown_method(self, method): self.data_api_client_patch.stop() super().teardown_method(method) @pytest.mark.parametrize("role,expected_code", [ ("admin", 403), ("admin-manager", 403), ("admin-ccs-category", 200), ("admin-ccs-sourcing", 200), ("admin-framework-manager", 200), ]) def test_outcomes_csv_download_permissions(self, role, expected_code): self.user_role = role response = self.client.get('/admin/direct-award/outcomes') actual_code = response.status_code assert actual_code == expected_code, "Unexpected response {} for role {}".format(actual_code, role) def test_outcomes_csv_download_content(self): self.user_role = 'admin-ccs-sourcing' find_direct_award_projects_result = { "links": { "self": "http://localhost:5000/direct-award/projects?latest-first=1&user-id=19175" }, "meta": { "total": 20 }, "projects": [ { "active": True, "createdAt": "2018-06-22T10:41:31.281853Z", "downloadedAt": None, "id": 731851428862851, "lockedAt": None, "name": "gfgffd", "outcome": { "result": "cancelled" }, "users": [ { "active": True, "emailAddress": "[email protected]", "id": 123, "name": "A Buyer", "role": "buyer" } ] }, { "active": True, "createdAt": "2018-06-19T13:36:37.557144Z", "downloadedAt": "2018-06-19T13:37:30.849304Z", "id": 272774709812396, "lockedAt": "2018-06-19T13:37:03.176398Z", "name": "22", "outcome": { "award": { "awardValue": "1234.00", "awardingOrganisationName": "123321", "endDate": "2020-12-12", "startDate": "2002-12-12" }, "completed": True, "completedAt": "2018-06-19T13:37:59.713497Z", "id": 680306864633356, "result": "awarded", "resultOfDirectAward": { "archivedService": { "id": 266018, "service": { "id": "316684326093280" } }, "project": { "id": 272774709812396 }, "search": { "id": 3706 } } }, "users": [ { "active": True, "emailAddress": "[email protected]", "id": 123, "name": "A Buyer", "role": "buyer" } ] } ] } get_archived_service_result = { 'services': { 'supplierId': 266018, 'supplierName': 'Somerford Associates Limited', 'serviceName': 'testServiceName' } } self.data_api_client.get_archived_service.return_value = get_archived_service_result self.data_api_client.find_direct_award_projects.return_value = find_direct_award_projects_result response = self.client.get('/admin/direct-award/outcomes') assert response.status_code == 200 assert response.content_type == 'text/csv; charset=utf-8' response_data = str(response.data, 'utf-8').splitlines() # convert byte-string to string data = csv.reader(response_data) assert data # checks if CSV is valid rows = [] for row in data: rows.append(row) # checks that only awarded outcomes are shown assert len(rows) == 2 # checks headers assert rows[0] == [ 'ID', 'Name', 'Submitted at', 'Result', 'Award service ID', 'Award service name', 'Award supplier id', 'Award supplier name', 'Award value', 'Awarding organisation name', 'Award start date', 'Award end date', 'User id', 'User name', 'User email' ] # checks results assert rows[1] == [ '272774709812396', '22', '2018-06-19T13:37:59.713497Z', 'awarded', '316684326093280', 'testServiceName', '266018', 'Somerford Associates Limited', '1234.00', '123321', '2002-12-12', '2020-12-12', '123', 'A Buyer', '[email protected]' ] class TestDOSView(LoggedInApplicationTest): url = "/admin/digital-outcomes-and-specialists/outcomes" def setup_method(self, method): super().setup_method(method) self.data_api_client_patch = mock.patch('app.main.views.outcomes.data_api_client', autospec=True) self.data_api_client = self.data_api_client_patch.start() self.data_api_client.find_frameworks.return_value = {"frameworks": [ FrameworkStub( slug="digital-outcomes-and-specialists-4", status="live" ).response() ]} def teardown_method(self, method): self.data_api_client_patch.stop() super().teardown_method(method) @pytest.fixture(autouse=True) def s3(self): with mock.patch("app.main.views.outcomes.s3") as s3: bucket = s3.S3() bucket.get_signed_url.side_effect = \ lambda path: f"https://s3.example.com/{path}?signature=deadbeef" yield s3 @pytest.mark.parametrize("role,expected_code", [ ("admin", 403), ("admin-manager", 403), ("admin-ccs-category", 302), ("admin-ccs-sourcing", 302), ("admin-framework-manager", 302), ]) def test_download_permissions(self, role, expected_code): self.user_role = role response = self.client.get(self.url) actual_code = response.status_code assert actual_code == expected_code, "Unexpected response {} for role {}".format(actual_code, role) def test_redirects_to_assets_domain(self): self.user_role = "admin-ccs-category" response = self.client.get(self.url) assert response.status_code == 302 assert response.location \ == "https://assets.test.digitalmarketplace.service.gov.uk" \ "/digital-outcomes-and-specialists-4/reports/opportunity-data.csv" \ "?signature=deadbeef" @pytest.mark.parametrize("latest_dos_framework", ( "digital-outcomes-and-specialists-4", "digital-outcomes-and-specialists-5", )) def test_csv_is_for_latest_live_dos_framework(self, latest_dos_framework, s3): self.user_role = "admin-ccs-category" self.data_api_client.find_frameworks.return_value = {"frameworks": [ FrameworkStub( framework_live_at="2016-03-03 12:00:00", slug="digital-outcomes-and-specialists", status="expired" ).response(), FrameworkStub( framework_live_at="2018-10-01 10:58:09.43134", slug="digital-outcomes-and-specialists-3", status="live" ).response(), FrameworkStub( framework_live_at="2019-12-18 15:13:24.53636", slug=latest_dos_framework, status="live" ).response(), FrameworkStub( framework_live_at="2020-12-18 15:13:24.53636", slug="g-cloud-12", status="live" ).response(), ]} response = self.client.get(self.url) assert s3.S3().get_signed_url.call_args == mock.call( f"{latest_dos_framework}/reports/opportunity-data.csv" ) assert latest_dos_framework in response.location
mit
-5,713,912,630,763,795,000
36.924686
107
0.492056
false
dc3-plaso/dfvfs
tests/path/os_path_spec.py
1
1764
#!/usr/bin/python # -*- coding: utf-8 -*- """Tests for the operating system path specification implementation.""" import platform import unittest from dfvfs.path import os_path_spec from tests.path import test_lib class OSPathSpecTest(test_lib.PathSpecTestCase): """Tests for the operating system path specification implementation.""" def testInitialize(self): """Tests the path specification initialization.""" if platform.system() == u'Windows': test_location = u'C:\\test' else: test_location = u'/test' path_spec = os_path_spec.OSPathSpec(location=test_location) self.assertIsNotNone(path_spec) with self.assertRaises(ValueError): _ = os_path_spec.OSPathSpec( location=test_location, parent=self._path_spec) with self.assertRaises(ValueError): _ = os_path_spec.OSPathSpec(location=test_location, bogus=u'BOGUS') def testComparable(self): """Tests the path specification comparable property.""" if platform.system() == u'Windows': test_location = u'C:\\test' else: test_location = u'/test' path_spec = os_path_spec.OSPathSpec(location=test_location) self.assertIsNotNone(path_spec) expected_comparable = u'\n'.join([ u'type: OS, location: {0:s}'.format(test_location), u'']) self.assertEqual(path_spec.comparable, expected_comparable) def testIsSystemLevel(self): """Tests the IsSystemLevel function.""" if platform.system() == u'Windows': test_location = u'C:\\test' else: test_location = u'/test' path_spec = os_path_spec.OSPathSpec(location=test_location) self.assertIsNotNone(path_spec) self.assertTrue(path_spec.IsSystemLevel()) if __name__ == '__main__': unittest.main()
apache-2.0
-6,657,228,756,776,693,000
26.138462
73
0.674036
false
georgeouzou/survgr
transform/migrations/0001_initial.py
1
1205
# -*- coding: utf-8 -*- # Generated by Django 1.9 on 2016-01-22 14:55 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Hattblock', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=25)), ('center_lon', models.FloatField()), ('center_lat', models.FloatField()), ('geometry', models.CharField(max_length=255)), ], ), migrations.CreateModel( name='OKXECoefficient', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('type', models.CharField(max_length=3)), ('value', models.FloatField()), ('block', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='transform.Hattblock')), ], ), ]
mit
4,114,823,400,723,477,000
33.428571
116
0.561826
false
fazalmajid/temboz
tembozapp/normalize.py
1
32692
# -*- coding: iso-8859-1 -*- from __future__ import print_function import sys, time, re, codecs, string, traceback, socket, hashlib import unicodedata, requests, feedparser from . import param, transform, util, porter2 import bleach try: import html.entities as htmlentitydefs except ImportError: import htmlentitydefs # XXX TODO # # XXX normalize feed['title'] to quote &amp; &quot; # # XXX Many of these heuristics have probably been addressed by newer versions # XXX of feedparser.py #date_fmt = '%a, %d %b %Y %H:%M:%S %Z' date_fmt = '%Y-%m-%d %H:%M:%S' try: try: import ctranslitcodec as translitcodec except ImportError: import translitcodec def strip_diacritics(s): return translitcodec.short_encode(s)[0] except ImportError: # strip diacritics. Unicode normalization form D (NFD) maps letters with # diacritics into the base letter followed by a combining diacritic, all # we need to do is get rid of the combining diacritics # this probably does not work with exotic characters like # U+FDF2 (Arabic ligature Allah) def stripc(c): return unicodedata.normalize('NFD', c)[0] def strip_diacritics(s): return ''.join(map(stripc, s)) # XXX need a good way to support languages other than English and French stop_words = ['i', 't', 'am', 'no', 'do', 's', 'my', 'don', 'm', 'on', 'get', 'in', 'you', 'me', 'd', 've'] # list originally from: http://bll.epnet.com/help/ehost/Stop_Words.htm stop_words += ['a', 'the', 'of', 'and', 'that', 'for', 'by', 'as', 'be', 'or', 'this', 'then', 'we', 'which', 'with', 'at', 'from', 'under', 'such', 'there', 'other', 'if', 'is', 'it', 'can', 'now', 'an', 'to', 'but', 'upon', 'where', 'these', 'when', 'whether', 'also', 'than', 'after', 'within', 'before', 'because', 'without', 'however', 'therefore', 'between', 'those', 'since', 'into', 'out', 'some', 'about', 'accordingly', 'again', 'against', 'all', 'almost', 'already', 'although', 'always', 'among', 'any', 'anyone', 'apparently', 'are', 'arise', 'aside', 'away', 'became', 'become', 'becomes', 'been', 'being', 'both', 'briefly', 'came', 'cannot', 'certain', 'certainly', 'could', 'etc', 'does', 'done', 'during', 'each', 'either', 'else', 'ever', 'every', 'further', 'gave', 'gets', 'give', 'given', 'got', 'had', 'hardly', 'has', 'have', 'having', 'here', 'how', 'itself', 'just', 'keep', 'kept', 'largely', 'like', 'made', 'mainly', 'make', 'many', 'might', 'more', 'most', 'mostly', 'much', 'must', 'nearly', 'necessarily', 'neither', 'next', 'none', 'nor', 'normally', 'not', 'noted', 'often', 'only', 'our', 'put', 'owing', 'particularly', 'perhaps', 'please', 'potentially', 'predominantly', 'present', 'previously', 'primarily', 'probably', 'prompt', 'promptly', 'quickly', 'quite', 'rather', 'readily', 'really', 'recently', 'regarding', 'regardless', 'relatively', 'respectively', 'resulted', 'resulting', 'results', 'said', 'same', 'seem', 'seen', 'several', 'shall', 'should', 'show', 'showed', 'shown', 'shows', 'significantly', 'similar', 'similarly', 'slightly', 'so', 'sometime', 'somewhat', 'soon', 'specifically', 'strongly', 'substantially', 'successfully', 'sufficiently', 'their', 'theirs', 'them', 'they', 'though', 'through', 'throughout', 'too', 'toward', 'unless', 'until', 'use', 'used', 'using', 'usually', 'various', 'very', 'was', 'were', 'what', 'while', 'who', 'whose', 'why', 'widely', 'will', 'would', 'yet'] # French stop words # list originally from: http://www.up.univ-mrs.fr/veronis/data/antidico.txt # XXX it's not good practice to mix languages like this, we should use # XXX feed language metadata and track what language a content is written in # XXX but that would require significant data model changes dec = (lambda s: unicode(s, 'iso8859-15')) if sys.version < '3' else str stop_words += [strip_diacritics(dec(s)) for s in [ "a", "A", "à", "afin", "ah", "ai", "aie", "aient", "aies", "ailleurs", "ainsi", "ait", "alentour", "alias", "allais", "allaient", "allait", "allons", "allez", "alors", "Ap.", "Apr.", "après", "après-demain", "arrière", "as", "assez", "attendu", "au", "aucun", "aucune", "au-dedans", "au-dehors", "au-delà", "au-dessous", "au-dessus", "au-devant", "audit", "aujourd'", "aujourd'hui", "auparavant", "auprès", "auquel", "aura", "aurai", "auraient", "aurais", "aurait", "auras", "aurez", "auriez", "aurions", "aurons", "auront", "aussi", "aussitôt", "autant", "autour", "autre", "autrefois", "autres", "autrui", "aux", "auxdites", "auxdits", "auxquelles", "auxquels", "avaient", "avais", "avait", "avant", "avant-hier", "avec", "avez", "aviez", "avions", "avoir", "avons", "ayant", "ayez", "ayons", "B", "bah", "banco", "bé", "beaucoup", "ben", "bien", "bientôt", "bis", "bon", "ç'", "c.-à-d.", "Ca", "ça", "çà", "cahin-caha", "car", "ce", "-ce", "céans", "ceci", "cela", "celle", "celle-ci", "celle-là", "celles", "celles-ci", "celles-là", "celui", "celui-ci", "celui-là", "cent", "cents", "cependant", "certain", "certaine", "certaines", "certains", "certes", "ces", "c'est-à-dire", "cet", "cette", "ceux", "ceux-ci", "ceux-là", "cf.", "cg", "cgr", "chacun", "chacune", "chaque", "cher", "chez", "ci", "-ci", "ci-après", "ci-dessous", "ci-dessus", "cinq", "cinquante", "cinquante-cinq", "cinquante-deux", "cinquante-et-un", "cinquante-huit", "cinquante-neuf", "cinquante-quatre", "cinquante-sept", "cinquante-six", "cinquante-trois", "cl", "cm", "cm²", "combien", "comme", "comment", "contrario", "contre", "crescendo", "D", "d'", "d'abord", "d'accord", "d'affilée", "d'ailleurs", "dans", "d'après", "d'arrache-pied", "davantage", "de", "debout", "dedans", "dehors", "déjà", "delà", "demain", "d'emblée", "depuis", "derechef", "derrière", "des", "dès", "desdites", "desdits", "désormais", "desquelles", "desquels", "dessous", "dessus", "deux", "devant", "devers", "dg", "die", "différentes", "différents", "dire", "dis", "disent", "dit", "dito", "divers", "diverses", "dix", "dix-huit", "dix-neuf", "dix-sept", "dl", "dm", "donc", "dont", "dorénavant", "douze", "du", "dû", "dudit", "duquel", "durant", "E", "eh", "elle", "-elle", "elles", "-elles", "en", "'en", "-en", "encore", "enfin", "ensemble", "ensuite", "entre", "entre-temps", "envers", "environ", "es", "ès", "est", "et", "et/ou", "étaient", "étais", "était", "étant", "etc", "été", "êtes", "étiez", "étions", "être", "eu", "eue", "eues", "euh", "eûmes", "eurent", "eus", "eusse", "eussent", "eusses", "eussiez", "eussions", "eut", "eût", "eûtes", "eux", "exprès", "extenso", "extremis", "F", "facto", "fallait", "faire", "fais", "faisais", "faisait", "faisaient", "faisons", "fait", "faites", "faudrait", "faut", "fi", "flac", "fors", "fort", "forte", "fortiori", "frais", "fûmes", "fur", "furent", "fus", "fusse", "fussent", "fusses", "fussiez", "fussions", "fut", "fût", "fûtes", "G", "GHz", "gr", "grosso", "guère", "H", "ha", "han", "haut", "hé", "hein", "hem", "heu", "hg", "hier", "hl", "hm", "hm³", "holà", "hop", "hormis", "hors", "hui", "huit", "hum", "I", "ibidem", "ici", "ici-bas", "idem", "il", "-il", "illico", "ils", "-ils", "ipso", "item", "J", "j'", "jadis", "jamais", "je", "-je", "jusqu'", "jusqu'à", "jusqu'au", "jusqu'aux", "jusque", "juste", "K", "kg", "km", "km²", "L", "l'", "la", "-la", "là", "-là", "là-bas", "là-dedans", "là-dehors", "là-derrière", "là-dessous", "là-dessus", "là-devant", "là-haut", "laquelle", "l'autre", "le", "-le", "lequel", "les", "-les", "lès", "lesquelles", "lesquels", "leur", "-leur", "leurs", "lez", "loin", "l'on", "longtemps", "lors", "lorsqu'", "lorsque", "lui", "-lui", "l'un", "l'une", "M", "m'", "m²", "m³", "ma", "maint", "mainte", "maintenant", "maintes", "maints", "mais", "mal", "malgré", "me", "même", "mêmes", "mes", "mg", "mgr", "MHz", "mieux", "mil", "mille", "milliards", "millions", "minima", "ml", "mm", "mm²", "modo", "moi", "-moi", "moins", "mon", "moult", "moyennant", "mt", "N", "n'", "naguère", "ne", "néanmoins", "neuf", "ni", "nº", "non", "nonante", "nonobstant", "nos", "notre", "nous", "-nous", "nul", "nulle", "O", "ô", "octante", "oh", "on", "-on", "ont", "onze", "or", "ou", "où", "ouais", "oui", "outre", "P", "par", "parbleu", "parce", "par-ci", "par-delà", "par-derrière", "par-dessous", "par-dessus", "par-devant", "parfois", "par-là", "parmi", "partout", "pas", "passé", "passim", "pendant", "personne", "petto", "peu", "peut", "peuvent", "peux", "peut-être", "pis", "plus", "plusieurs", "plutôt", "point", "posteriori", "pour", "pourquoi", "pourtant", "préalable", "près", "presqu'", "presque", "primo", "priori", "prou", "pu", "puis", "puisqu'", "puisque", "Q", "qu'", "qua", "quand", "quarante", "quarante-cinq", "quarante-deux", "quarante-et-un", "quarante-huit", "quarante-neuf", "quarante-quatre", "quarante-sept", "quarante-six", "quarante-trois", "quasi", "quatorze", "quatre", "quatre-vingt", "quatre-vingt-cinq", "quatre-vingt-deux", "quatre-vingt-dix", "quatre-vingt-dix-huit", "quatre-vingt-dix-neuf", "quatre-vingt-dix-sept", "quatre-vingt-douze", "quatre-vingt-huit", "quatre-vingt-neuf", "quatre-vingt-onze", "quatre-vingt-quatorze", "quatre-vingt-quatre", "quatre-vingt-quinze", "quatre-vingts", "quatre-vingt-seize", "quatre-vingt-sept", "quatre-vingt-six", "quatre-vingt-treize", "quatre-vingt-trois", "quatre-vingt-un", "quatre-vingt-une", "que", "quel", "quelle", "quelles", "quelqu'", "quelque", "quelquefois", "quelques", "quelques-unes", "quelques-uns", "quelqu'un", "quelqu'une", "quels", "qui", "quiconque", "quinze", "quoi", "quoiqu'", "quoique", "R", "revoici", "revoilà", "rien", "S", "s'", "sa", "sans", "sauf", "se", "secundo", "seize", "selon", "sensu", "sept", "septante", "sera", "serai", "seraient", "serais", "serait", "seras", "serez", "seriez", "serions", "serons", "seront", "ses", "si", "sic", "sine", "sinon", "sitôt", "situ", "six", "soi", "soient", "sois", "soit", "soixante", "soixante-cinq", "soixante-deux", "soixante-dix", "soixante-dix-huit", "soixante-dix-neuf", "soixante-dix-sept", "soixante-douze", "soixante-et-onze", "soixante-et-un", "soixante-et-une", "soixante-huit", "soixante-neuf", "soixante-quatorze", "soixante-quatre", "soixante-quinze", "soixante-seize", "soixante-sept", "soixante-six", "soixante-treize", "soixante-trois", "sommes", "son", "sont", "soudain", "sous", "souvent", "soyez", "soyons", "stricto", "suis", "sur", "sur-le-champ", "surtout", "sus", "T", "-t", "t'", "ta", "tacatac", "tant", "tantôt", "tard", "te", "tel", "telle", "telles", "tels", "ter", "tes", "toi", "-toi", "ton", "tôt", "toujours", "tous", "tout", "toute", "toutefois", "toutes", "treize", "trente", "trente-cinq", "trente-deux", "trente-et-un", "trente-huit", "trente-neuf", "trente-quatre", "trente-sept", "trente-six", "trente-trois", "très", "trois", "trop", "tu", "-tu", "U", "un", "une", "unes", "uns", "USD", "V", "va", "vais", "vas", "vers", "veut", "veux", "via", "vice-versa", "vingt", "vingt-cinq", "vingt-deux", "vingt-huit", "vingt-neuf", "vingt-quatre", "vingt-sept", "vingt-six", "vingt-trois", "vis-à-vis", "vite", "vitro", "vivo", "voici", "voilà", "voire", "volontiers", "vos", "votre", "vous", "-vous", "W", "X", "y", "-y", "Z", "zéro"]] stop_words = set(stop_words) # translate to lower case, normalize whitespace # for ease of filtering # this needs to be a mapping as Unicode strings do not support traditional # str.translate with a 256-length string lc_map = {} punct_map = {} for c in string.whitespace: lc_map[ord(c)] = 32 del lc_map[32] for c in string.punctuation + '\'\xab\xbb': punct_map[ord(c)] = 32 punct_map[0x2019] = "'" # decode HTML entities with known Unicode equivalents ent_re = re.compile(r'\&([^;]*);') def ent_sub(m): ent = m.groups()[0] if ent in htmlentitydefs.name2codepoint: return chr(htmlentitydefs.name2codepoint[ent]) if ent.startswith('#'): if ent.lower().startswith('#x'): codepoint = int('0x' + ent[2:], 16) else: try: codepoint = int(ent[1:]) except ValueError: return ent if codepoint > 0 and codepoint < sys.maxunicode: return chr(codepoint) # fallback - leave as-is return '&%s;' % ent def decode_entities(s): return ent_re.sub(ent_sub, s) # XXX need to normalize for HTML entities as well def lower(s): """Turn a string lower-case, including stripping accents""" return strip_diacritics(decode_entities(s)).translate(lc_map).lower() # XXX this implementation is hopefully correct, but inefficient # XXX we should be able to replace it with a finite state automaton in C # XXX for better performance # tested with u=u'\xe9sop\xe9sopfoo\xe9sop' and unicodedata.normalize('NFD', u) def replace_first(s, pat, mark_begin, mark_end): """Case-insensitive replacement of the 1st occurrence of pat in s by repl""" lc = lower(s) pat = lower(pat) start = lc.find(pat) if start == -1: return s else: # (c)translitcodec does more than simply lowercasing, so we will need # to use bisection to find where in the untransliterated string the # pattern can be found if lower(s[start:]).find(pat) == -1: # As a fast-path, use the position in the transliterated string as # an initial guess of where to start, but in this case it did not work start = 0 end = len(s) while lower(s[start:]).find(pat) > 0: if start == end: # XXX still can't find it, this shouldn't happen return s mid = (start + end + 1) // 2 if lower(s[mid:]).find(pat) >= 0: start = mid else: end = mid # now we have the start, find the end end = start + len(pat) if lower(s[start:end]) != pat: end = start # the pattern may not be equal, e.g. searching for 'GB' in '£' that # expands to 'gbp' while not lower(s[start:end]).startswith(pat): end += 1 return s[:start] + mark_begin + s[start:end] + mark_end + s[end:] strip_tags_re = re.compile('<[^>]*>') def get_words(s): return set([ word for word in lower(str(strip_tags_re.sub('', str(s))) ).translate(punct_map).split() if word not in stop_words]) def stem(words): return {porter2.stem(word) for word in words} ######################################################################## # HTML tag balancing logic # # from the HTML4 loose DTD http://www.w3.org/TR/html4/loose.dtd fontstyle = ('b', 'big', 'i', 's', 'small', 'strike', 'tt', 'u') phrase = ('abbr', 'acronym', 'cite', 'code', 'dfn', 'em', 'kbd', 'samp', 'strong', 'var') heading = ('h1', 'h2', 'h3', 'h4', 'h5', 'h6') html4_elts = ('a', 'address', 'applet', 'area', 'base', 'basefont', 'bdo', 'blockquote', 'body', 'br', 'button', 'caption', 'center', 'col', 'colgroup', 'dd', 'del', 'dir', 'div', 'dl', 'dt', 'fieldset', 'font', 'form', 'frame', 'frameset', 'head', 'hr', 'html', 'iframe', 'img', 'input', 'ins', 'isindex', 'label', 'legend', 'li', 'link', 'map', 'menu', 'meta', 'noframes', 'noscript', 'object', 'ol', 'optgroup', 'option', 'p', 'param', 'pre', 'q', 'script', 'select', 'span', 'style', 'sub', 'sup', 'table', 'tbody', 'td', 'textarea', 'tfoot', 'th', 'thead', 'title', 'tr', 'ul') + fontstyle + phrase + heading inline_elts = ('a', 'abbr', 'acronym', 'address', 'bdo', 'caption', 'cite', 'code', 'dfn', 'dt', 'em', 'font', 'i', 'iframe', 'kbd', 'label', 'legend', 'p', 'pre', 'q', 's', 'samp', 'small', 'span', 'strike', 'strong', 'sub', 'sup', 'tt', 'u', 'var') + fontstyle + phrase + heading # strictly speaking the closing '</p> tag is optional, but let's close it # since it is so common closing = ('a', 'address', 'applet', 'bdo', 'blockquote', 'button', 'caption', 'center', 'del', 'dir', 'div', 'dl', 'fieldset', 'font', 'form', 'frameset', 'iframe', 'ins', 'label', 'legend', 'map', 'menu', 'noframes', 'noscript', 'object', 'ol', 'optgroup', 'pre', 'q', 'script', 'select', 'span', 'style', 'sub', 'sup', 'table', 'textarea', 'title', 'ul') + fontstyle + phrase + heading + ('p',) # <!ENTITY % block block = ('address', 'blockquote', 'center', 'dir', 'div', 'dl', 'fieldset', 'form', 'hr', 'isindex', 'menu', 'noframes', 'noscript', 'ol', 'p', 'pre', 'table', 'ul') + heading # for XSS attacks, as feedparser is not completely immune banned = ('script', 'applet', 'style') # speed up things a bit block = set(block) closing = set(closing) banned = set(banned) acceptable_elements = set([ 'a', 'abbr', 'acronym', 'address', 'area', 'article', 'aside', 'audio', 'b', 'big', 'blockquote', 'br', 'button', 'canvas', 'caption', 'center', 'cite', 'code', 'col', 'colgroup', 'command', 'datagrid', 'datalist', 'dd', 'del', 'details', 'dfn', 'dialog', 'dir', 'div', 'dl', 'dt', 'em', 'event-source', 'fieldset', 'figcaption', 'figure', 'footer', 'font', 'form', 'header', 'h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'hr', 'i', 'img', 'input', 'ins', 'keygen', 'kbd', 'label', 'legend', 'li', 'm', 'map', 'menu', 'meter', 'multicol', 'nav', 'nextid', 'ol', 'output', 'optgroup', 'option', 'p', 'pre', 'progress', 'q', 's', 'samp', 'section', 'select', 'small', 'sound', 'source', 'spacer', 'span', 'strike', 'strong', 'sub', 'sup', 'table', 'tbody', 'td', 'textarea', 'time', 'tfoot', 'th', 'thead', 'tr', 'tt', 'u', 'ul', 'var', 'video', 'noscript' ]) acceptable_attributes = [ 'abbr', 'accept', 'accept-charset', 'accesskey', 'action', 'align', 'alt', 'autocomplete', 'autofocus', 'axis', 'background', 'balance', 'bgcolor', 'bgproperties', 'border', 'bordercolor', 'bordercolordark', 'bordercolorlight', 'bottompadding', 'cellpadding', 'cellspacing', 'ch', 'challenge', 'char', 'charoff', 'choff', 'charset', 'checked', 'cite', 'class', 'clear', 'color', 'cols', 'colspan', 'compact', 'contenteditable', 'controls', 'coords', 'data', 'datafld', 'datapagesize', 'datasrc', 'datetime', 'default', 'delay', 'dir', 'disabled', 'draggable', 'dynsrc', 'enctype', 'end', 'face', 'for', 'form', 'frame', 'galleryimg', 'gutter', 'headers', 'height', 'hidefocus', 'hidden', 'high', 'href', 'hreflang', 'hspace', 'icon', 'id', 'inputmode', 'ismap', 'keytype', 'label', 'leftspacing', 'lang', 'list', 'longdesc', 'loop', 'loopcount', 'loopend', 'loopstart', 'low', 'lowsrc', 'max', 'maxlength', 'media', 'method', 'min', 'multiple', 'name', 'nohref', 'noshade', 'nowrap', 'open', 'optimum', 'pattern', 'ping', 'point-size', 'poster', 'pqg', 'preload', 'prompt', 'radiogroup', 'readonly', 'rel', 'repeat-max', 'repeat-min', 'replace', 'required', 'rev', 'rightspacing', 'rows', 'rowspan', 'rules', 'scope', 'selected', 'shape', 'size', 'span', 'src', 'start', 'step', 'summary', 'suppress', 'tabindex', 'target', 'template', 'title', 'toppadding', 'type', 'unselectable', 'usemap', 'urn', 'valign', 'value', 'variable', 'volume', 'vspace', 'vrml', 'width', 'wrap' ] acceptable_css_properties = [ 'azimuth', 'background-color', 'border-bottom-color', 'border-collapse', 'border-color', 'border-left-color', 'border-right-color', 'border-top-color', 'clear', 'color', 'cursor', 'direction', 'display', 'elevation', 'float', 'font', 'font-family', 'font-size', 'font-style', 'font-variant', 'font-weight', 'height', 'letter-spacing', 'line-height', 'overflow', 'pause', 'pause-after', 'pause-before', 'pitch', 'pitch-range', 'richness', 'speak', 'speak-header', 'speak-numeral', 'speak-punctuation', 'speech-rate', 'stress', 'text-align', 'text-decoration', 'text-indent', 'unicode-bidi', 'vertical-align', 'voice-family', 'volume', 'white-space', 'width' ] def sanitize_text(text): """Sanitize text fields like title or feed description for XSS""" return bleach.clean( text, tags=[], attributes=[], styles=[], strip=True ) tag_re = re.compile(r'(<>|<[^!].*?>|<!\[CDATA\[|\]\]>|<!--.*?-->|<[!]>)', re.DOTALL | re.MULTILINE) def balance(html, limit_words=None, ellipsis=' ...'): # we cannot trust feedparser to sanitize if not limit_words: #return html5lib.serialize(html5lib.parse(html)) return bleach.clean( html, tags=acceptable_elements, attributes=acceptable_attributes, styles=acceptable_css_properties, strip=True ) # the legacy balancing logic is redundant with Bleach's, # but this is seldom used word_count = 0 tokens = tag_re.split(html) out = [] stack = [] for token in tokens: if not token.startswith('<'): if limit_words and word_count > limit_words: break words = token.split() word_count += len(words) if limit_words and word_count > limit_words: crop = limit_words - word_count out.append(' '.join(words[:crop]) + ellipsis) else: out.append(token) continue if token.startswith('<!'): continue if token == ']]>': continue if not token.endswith('>'): continue # invalid element = token[1:-1].split()[0].lower() if not element: continue # invalid if element in banned: element = 'pre' token = '<pre>' if element.startswith('/'): element = element[1:] if element in banned: element = 'pre' token = '</pre>' if element in stack: top = None while stack and top != element: top = stack.pop() out.append('</%s>' % top) continue else: continue if element in block and stack and stack[-1] not in block: # close previous block if any for i in range(len(stack) - 1, -1, -1): if stack[i] in block: break stack, previous_block = stack[:i], stack[i:] previous_block.reverse() for tag in previous_block: out.append('</%s>' % tag) if element in closing and not token.endswith('/>'): stack.append(element) out.append(token) # flush the stack out.extend(['</%s>' % element for element in reversed(stack)]) html = ''.join(out) return bleach.clean( html, tags=acceptable_elements, attributes=acceptable_attributes, styles=acceptable_css_properties, strip=True ) ######################################################################## def normalize_all(f): normalize_feed(f) for item in f.entries: normalize(item, f) def normalize_feed(f): if 'description' not in f['channel']: f['channel']['description'] = f['channel'].get('title', '') f['channel']['description'] = sanitize_text(f['channel']['description']) if 'modified' in f and type(f['modified']) == str: try: f['modified'] = time.strptime(f['modified'], '%a, %d %b %Y %H:%M:%S GMT') except ValueError: f['modified'] = time.strptime(f['modified'], '%a, %d %b %Y %H:%M:%S +0000') # Often, broken RSS writers will not handle daylight savings time correctly # and use a timezone that is off by one hour. For instance, in the US/Pacific # time zone: # February 3, 2004, 5:30PM is 2004-02-03T17:30:00-08:00 (standard time) # August 3, 2004, 5:30PM US/Pacific is 2004-08-03T17:30:00-07:00 (DST) # but broken implementations will incorrectly write: # 2004-08-03T17:30:00-08:00 in the second case # There is no real good way to ward against this, but if the created or # modified date is in the future, we are clearly in this situation and # substract one hour to correct for this bug def fix_date(date_tuple): if not date_tuple: return date_tuple if date_tuple > time.gmtime(): # feedparser's parsed date tuple has no DST indication, we need to force it # because there is no UTC equivalent of mktime() date_tuple = date_tuple[:-1] + (-1,) date_tuple = time.localtime(time.mktime(date_tuple) - 3600) # if it is still in the future, the implementation is hopelessly broken, # truncate it to the present if date_tuple > time.gmtime(): return time.gmtime() else: return date_tuple else: return date_tuple # why doesn't feedparser do these basic normalizations? def basic(f, feed_xml): if 'url' not in f: f['url'] = feed_xml # CVS versions of feedparser are not throwing exceptions as they should # see: # http://sourceforge.net/tracker/index.php?func=detail&aid=1379172&group_id=112328&atid=661937 if not f.feed or ('link' not in f.feed or 'title' not in f.feed): # some feeds have multiple links, one for self and one for PuSH if f.feed and 'link' not in f.feed and 'links' in f.feed: try: for l in f.feed['links']: if l['rel'] == 'self': f.feed['link'] = l['href'] except KeyError: pass if 'title' in f.feed: f.feed['title'] = sanitize_text(f.feed['title']) def dereference(url, seen=None, level=0): """Recursively dereference a URL""" # this set is used to detect redirection loops if seen is None: seen = set([url]) else: seen.add(url) # stop recursion if it is too deep if level > 16: return url try: r = requests.get(url, allow_redirects=False, timeout=param.http_timeout) if not r.is_redirect: return url else: # break a redirection loop if it occurs redir = r.headers.get('Location') if True not in [redir.startswith(p) for p in ['http://', 'https://', 'ftp://']]: return url if redir in seen: return url # some servers redirect to Unicode URLs, which are not legal try: str(redir) except UnicodeDecodeError: return url # there might be several levels of redirection return dereference(redir, seen, level + 1) except (requests.exceptions.RequestException, ValueError, socket.error): return url except: util.print_stack() return url url_re = re.compile('(?:href|src)="([^"]*)"', re.IGNORECASE) def normalize(item, f, run_filters=True): # get rid of RDF lossage... for key in ['title', 'link', 'created', 'modified', 'author', 'content', 'content_encoded', 'description']: if type(item.get(key)) == list: if len(item[key]) == 1: item[key] = item[key][0] else: candidate = [i for i in item[key] if i.get('type') == 'text/html'] if len(candidate) > 1 and key == 'content': candidate = sorted(candidate, key=lambda i: len(i.get('value', '')), reverse=True)[:1] if len(candidate) == 1: item[key] = candidate[0] else: # XXX not really sure how to handle these cases print('E' * 16, 'ambiguous RDF', key, item[key], file=param.log) item[key] = item[key][0] if isinstance(item.get(key), dict) and 'value' in item[key]: item[key] = item[key]['value'] ######################################################################## # title if 'title' not in item or not item['title'].strip(): item['title'] = 'Untitled' item['title'] = sanitize_text(item['title']) item['title_lc'] = lower(item['title']) item['title_words_exact'] = get_words(item['title_lc']) item['title_words'] = stem(item['title_words_exact']) ######################################################################## # link # # The RSS 2.0 specification allows items not to have a link if the entry # is complete in itself # that said this is almost always spurious, so we filter it below if 'link' not in item: item['link'] = f['channel']['link'] # We have to be careful not to assign a default URL as the GUID # otherwise only one item will ever be recorded if 'id' not in item: item['id'] = 'HASH_CONTENT' item['RUNT'] = True ######################################################################## # GUID if 'id' not in item: item['id'] = item['link'] ######################################################################## # creator if 'author' not in item or item['author'] == 'Unknown': item['author'] = 'Unknown' if 'author' in f['channel']: item['author'] = f['channel']['author'] item['author'] = sanitize_text(item['author']) ######################################################################## # created amd modified dates if 'modified' not in item: item['modified'] = f['channel'].get('modified') # created - use modified if not available if 'created' not in item: if 'modified_parsed' in item: created = item['modified_parsed'] else: created = None else: created = item['created_parsed'] if not created: # XXX use HTTP last-modified date here created = time.gmtime() # feeds that do not have timestamps cannot be garbage-collected # XXX need to find a better heuristic, as high-volume sites such as # XXX The Guardian, CNET.com or Salon.com lack item-level timestamps f['oldest'] = '1970-01-01 00:00:00' created = fix_date(created) item['created'] = time.strftime(date_fmt, created) # keep track of the oldest item still in the feed file if 'oldest' not in f: f['oldest'] = '9999-99-99 99:99:99' if item['created'] < f['oldest']: f['oldest'] = item['created'] # finish modified date if 'modified_parsed' in item and item['modified_parsed']: modified = fix_date(item['modified_parsed']) # add a fudge factor time window within which modifications are not # counted as such, 10 minutes here if not modified or abs(time.mktime(modified) - time.mktime(created)) < 600: item['modified'] = None else: item['modified'] = time.strftime(date_fmt, modified) else: item['modified'] = None ######################################################################## # content if 'content' in item: content = item['content'] elif 'content_encoded' in item: content = item['content_encoded'] elif 'description' in item: content = item['description'] else: content = '<a href="' + item['link'] + '">' + item['title'] + '</a>' if not content: content = '<a href="' + item['link'] + '">' + item['title'] + '</a>' # strip embedded NULs as a defensive measure content = content.replace('\0', '') # apply ad filters and other degunking to content old_content = None while old_content != content: old_content = content try: for filter in transform.filter_list: content = filter.apply(content, f, item) except: util.print_stack(black_list=['item']) # balance tags like <b>...</b> and sanitize content = balance(content) content_lc = lower(content) # the content might have invalid 8-bit characters. # Heuristic suggested by Georg Bauer if type(content) != str: try: content = content.decode('utf-8') except UnicodeError: content = content.decode('iso-8859-1') # item['content'] = content # we recalculate this as content may have changed due to tag rebalancing, etc item['content_lc'] = lower(content) item['content_words_exact'] = get_words(item['content_lc']) item['content_words'] = stem(item['content_words_exact']) item['union_lc'] = item['title_lc'] + '\n' + item['content_lc'] item['union_words'] = item['title_words'].union(item['content_words']) item['urls'] = url_re.findall(content) ######################################################################## # categories/tags # we used 'category' before, but 'category' and 'categories' are # intercepted by feedparser.FeedParserDict.__getitemm__ and treated as # special case if 'tags' in item and type(item['tags']) == list: item['item_tags'] = set([lower(sanitize_text(t['term'])) for t in item['tags']]) else: item['item_tags'] = [] ######################################################################## # map unicode # for key in ['title', 'link', 'created', 'modified', 'author', 'content']: # if type(item.get(key)) == str: # item[key] = item[key].encode('ascii', 'xmlcharrefreplace') # hash the content as the GUID if required if item['id'] == 'HASH_CONTENT': item['id']= hashlib.md5( (item['title'] + item['content']).encode('utf-8')).hexdigest() return item def escape_xml(s): """Escape entities for a XML target""" try: s = s.decode('utf-8') except: pass return s.replace('&', '&amp;').replace("'", "&apos;").replace('"', '&quot;').replace('<', '&lt;').replace('>', '&gt;').encode('ascii', 'xmlcharrefreplace').decode('ascii')
mit
-4,501,585,619,127,446,500
43.844993
173
0.580693
false
a-krebs/finances
finances/finances/urls.py
1
1583
# Copyright (C) 2012 Aaron Krebs [email protected] # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/> from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() # development serving of static files # TODO: before deployment, server static files properly from django.contrib.staticfiles.urls import staticfiles_urlpatterns urlpatterns = patterns('', # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), url(r'^accounts/', include('accounts.urls', namespace='accounts')), url(r'^budgets/', include('budgets.urls', namespace='budgets')), url(r'^register/', include('django_registration.urls', namespace='registration')), ) # TODO: remove before deployment urlpatterns += staticfiles_urlpatterns()
gpl-3.0
1,204,365,858,756,522,800
40.657895
86
0.749842
false
XD-embedded/xd-build-core
tests/recipe_file_test.py
1
4304
from xd.build.core.recipe_file import * from case import * import unittest import io class tests(unittest.case.TestCase): def test_split_no_version(self): name, version = RecipeFile.split_name_and_version('foo') self.assertEqual(name, 'foo') self.assertEqual(str(version), '') def test_split_version_1(self): name, version = RecipeFile.split_name_and_version('foo_4.2') self.assertEqual(name, 'foo') self.assertEqual(str(version), '4.2') def test_split_version_2(self): name, version = RecipeFile.split_name_and_version('foo_4.2.xd') self.assertEqual(name, 'foo') self.assertEqual(str(version), '4.2') def test_split_bad_version_1(self): with self.assertRaises(InvalidRecipeName): RecipeFile.split_name_and_version('') def test_split_bad_version_2(self): with self.assertRaises(InvalidRecipeName): RecipeFile.split_name_and_version('foo_bar_4.2') def test_with_path_1(self): recipe_file = RecipeFile('/path/to/something/foo.xd') self.assertEqual(recipe_file.name, 'foo') self.assertEqual(str(recipe_file.version), '') def test_with_path_2(self): recipe_file = RecipeFile('/some/path/bar_31.7.xd') self.assertEqual(recipe_file.name, 'bar') self.assertEqual(str(recipe_file.version), '31.7') def test_with_odd_path_1(self): with self.assertRaises(InvalidRecipeFilename): RecipeFile('/some/path/.xd') def test_bad_filename_1(self): with self.assertRaises(InvalidRecipeFilename): RecipeFile('/tmp/foo.bar') def test_bad_filename_2(self): with self.assertRaises(InvalidRecipeFilename): RecipeFile('/tmp/foo') def test_badd_filename_3(self): with self.assertRaises(InvalidRecipeFilename): RecipeFile('/some/path/.xd') def test_badd_filename_4(self): with self.assertRaises(InvalidRecipeFilename): RecipeFile('/some/path/foo_bar_1.xd') def test_with_odd_name(self): recipe_file = RecipeFile('/some/path/bar.93-1_4.xd') self.assertEqual(recipe_file.name, 'bar.93-1') self.assertEqual(str(recipe_file.version), '4') def test_with_odd_version_1(self): recipe_file = RecipeFile('/some/path/bar_4.2.1rc3.1-1.xd') self.assertEqual(recipe_file.name, 'bar') self.assertEqual(str(recipe_file.version), '4.2.1rc3.1-1') def test_with_odd_version_2(self): recipe_file = RecipeFile('/some/path/bar_89.23~build-189.xd') self.assertEqual(recipe_file.name, 'bar') self.assertEqual(str(recipe_file.version), '89.23~build-189') def test_str_1(self): recipe_file = RecipeFile('/tmp/foo.xd') self.assertEqual(str(recipe_file), 'foo') def test_str_2(self): recipe_file = RecipeFile('/tmp/foo_1.89.xd') self.assertEqual(str(recipe_file), 'foo_1.89') def test_repr(self): recipe_file = RecipeFile('/tmp/foo_1.89.xd') self.assertEqual(repr(recipe_file), "RecipeFile('/tmp/foo_1.89.xd')") def test_eq_1(self): recipe_file_a = RecipeFile('/tmp/foo_1.89.xd') recipe_file_b = RecipeFile('/tmp/foo_1.89.xd') self.assertEqual(recipe_file_a, recipe_file_b) def test_eq_2(self): recipe_file_a = RecipeFile('/tmp/foo_1.89.xd') recipe_file_b = RecipeFile('/tmp/foo_1.90.xd') self.assertNotEqual(recipe_file_a, recipe_file_b) def test_eq_3(self): recipe_file_a = RecipeFile('/tmp/foo_1.89.xd') recipe_file_b = '/tmp/foo_1.89.xd' self.assertNotEqual(recipe_file_a, recipe_file_b) class parse_tests(TestCase): def test_parse_1(self): with open('foobar.xd', 'w') as f: f.write('FOO="foo"\n') recipe_file = RecipeFile('foobar.xd') d = recipe_file.parse() self.assertEqual(len(d), 1) self.assertEqual(d['FOO'].get(), 'foo') def test_dump_1(self): with open('foobar.xd', 'w') as f: f.write('FOO="foo"\n') recipe_file = RecipeFile('foobar.xd') recipe_file.parse() stream = io.StringIO() recipe_file.dump(stream=stream) self.assertRegex("FOO='foo'\n", stream.getvalue())
mit
5,633,485,530,337,033,000
34.278689
77
0.619424
false
oztalha/weatherforecast
weather.py
1
14123
""" author Talha Oz this is the main code implemented for cs780 class project """ # -*- coding: utf-8 -*- import pandas as p import numpy as np from sklearn.feature_extraction.text import TfidfVectorizer import re from sklearn import cross_validation import preprocessing as pre from sklearn.naive_bayes import MultinomialNB from sklearn.decomposition import PCA from sklearn.multiclass import OneVsRestClassifier from sklearn import linear_model import matplotlib.pyplot as plt from variableNames import * import scipy.sparse from sklearn.linear_model import SGDRegressor from sklearn.linear_model import SGDClassifier from sklearn.neighbors import KNeighborsRegressor from sklearn.feature_extraction.text import CountVectorizer from sklearn.feature_selection import SelectKBest, chi2 """ from sklearn.svm import SVC from sklearn.decomposition import TruncatedSVD from sklearn.ensemble import RandomForestClassifier from sklearn.ensemble import ExtraTreesClassifier from sklearn.linear_model import Perceptron from sklearn import svm from sklearn.preprocessing import LabelBinarizer import matplotlib.pyplot as plt from scipy import sparse from sklearn import decomposition from sklearn.ensemble import ExtraTreesRegressor import sklearn.decomposition as deco import argparse from sklearn.svm import SVC %autoreload 2 """ def plotClasses(y): """ each class is counted by its weight, not # of nonzero occurrences """ fig = plt.figure() ax = plt.subplot(1,1,1) x1 = range(y.shape[1]) y1 = [sum(y[:,a]) for a in range(y.shape[1])] width = 0.8 labels = "s1,s2,s3,s4,s5,w1,w2,w3,w4,k1,k2,k3,k4,k5,k6,k7,k8,k9,k10,k11,k12,k13,k14,k15".split(',') if y.shape[1] == 5: labels = labels[:5] elif y.shape[1] == 4: labels = labels[5:9] else: labels = labels[9:] plt.xticks(np.arange(y.shape[1])+width/2,labels) legendkeys = tuple([k for k,v in legend.items() if k in labels]) legendvalues= tuple([v for k,v in legend.items() if k in labels]) [ax.bar(X,Y,width=width,label=k+' '+v) for X,Y,k,v in zip(x1,y1,legendkeys,legendvalues)] # Shink current axis by 20% box = ax.get_position() ax.set_position([box.x0, box.y0, box.width * 0.8, box.height]) # Put a legend to the right of the current axis ax.legend(loc='center left', bbox_to_anchor=(1, 0.5)) #ax.legend(('1','2'),('1a','2a'))#legendkeys,legendvalues plt.show() def vectorize(y,train,t2,model,kbest=0,is_tfidf=True,is_sparse=True,max_features=None,is_cv=False,perlabel=False,min_df=5,is_nominal=False,is_normal=False,is_LSA=False,scale=False): if is_cv: X_train, X_test, y_train, y_true = cross_validation.train_test_split(train['tweet'], y, test_size=.20, random_state = 0) # X_train1, X_test1, y_train, y_true = cross_validation.train_test_split(train['tweet'], y, test_size=.20, random_state = 0) # X_train2, X_test2, y_train, y_true = cross_validation.train_test_split(train['state'], y, test_size=.20, random_state = 0) # X_train3, X_test3, y_train, y_true = cross_validation.train_test_split(train['location'], y, test_size=.20, random_state = 0) # X_train = np.hstack((X_train1,X_train2,X_train3)) # X_test = np.hstack((X_test1,X_test2,X_test3)) else: X_train = train['tweet'] X_test = t2['tweet'] y_train = y # if (y_train.shape[1] > 6): # model = linear_model.Ridge (alpha = 3.0, normalize = False) # if is_PCA: # modelPCA = PCA(n_components='mle') # model.fit(X_train) if is_tfidf: #tfidf = TfidfVectorizer(max_features=max_features, strip_accents='unicode', analyzer='word', smooth_idf=True,sublinear_tf=True,max_df=0.5,min_df=min_df,ngram_range=(1,2),use_idf=True) tfidf = TfidfVectorizer(max_features=max_features,strip_accents='unicode', analyzer='word', smooth_idf=True,sublinear_tf=True,max_df=0.5,min_df=5,ngram_range=(1,2),use_idf=True) #tfidf.fit(np.hstack((X_train,X_test))) #fitting on the whole data resulted in a worse mse score tfidf.fit(X_train) X_train = tfidf.transform(X_train) X_test = tfidf.transform(X_test) if is_LSA: LSA = TruncatedSVD(n_components=10000, algorithm='randomized', n_iter=5, random_state=0, tol=0.0) X_train = LSA.fit_transform(X_train) X_test = LSA.transform(X_test) else: vectorizer = CountVectorizer( binary = True ) X_train = vectorizer.fit_transform(X_train) X_test = vectorizer.transform(X_test) if is_nominal: if (y_train.shape[1] < 16): y_rest = y_train.copy() X_train_list = [] y_weight_list = [] y_train_list = [] for i in range(y_rest.shape[1]): X_train_list.append(X_train) # repeat X to matchup y_weight_list.append(np.apply_along_axis(lambda a: a.max(), 1, y_rest)) # get the maximum in y_rest y_train_list.append(np.apply_along_axis(lambda a: a.argmax(), 1, y_rest).astype(int)) # get the position of the maximum in y_rest y_rest = np.apply_along_axis(lambda a: [0 if i == a.argmax() else x for i,x in enumerate(a)], 1, y_rest) #set maximum to zero # y_weight = np.concatenate((y_weight, np.apply_along_axis(lambda a: a.max(), 1, y_rest))) # y_train = np.concatenate((y_train, np.apply_along_axis(lambda a: a.argmax(), 1, y_rest).astype(int))) # y_train = np.apply_along_axis(lambda a: [np.floor(x) if x != a.max() else 1 for x in a], 1, y_train).astype(bool) not_kind = True X_train = scipy.sparse.vstack(X_train_list) y_train = np.concatenate(y_train_list) y_weight = np.concatenate(y_weight_list) else: not_kind = False #y_train = np.apply_along_axis(lambda a: [np.floor(x) if i != a.argmax() else 1 for i,x in enumerate(a)], 1, y_train).astype(bool) #y_train = np.ceil(y_train).astype(bool) #y_weight = y_train.copy() if perlabel: test_prediction=[] for i in range(y_train.shape[1]): if is_nominal: model.fit(X_train,y_train[:,i]) #sample_weight=y_weight[:,i] pred = model.predict_proba(X_test) # pred = model.predict_log_proba(X_test) # for log in SGDRegressor print pred.shape test_prediction.append(pred) else: model.fit(X_train,y_train[:,i]) test_prediction.append(model.predict(X_test)) pred = np.array(test_prediction).T if kbest: ch2 = SelectKBest(chi2,kbest,k=1000) #yb = y_train yb = np.around(y_train).astype(bool) X_train = ch2.fit_transform(X_train, yb) X_test = ch2.transform(X_test) if not is_sparse: X_train = X_train.toarray() X_test = X_test.toarray() #nmf = decomposition.NMF(n_components=y_train.shape[1]).fit(tfidf) #cca = CCA(n_components=100) #X_train = cca.fit_transform(X_train) #X_test = cca.transform(X_test) if not perlabel: if is_nominal and not_kind: model.fit(X_train, y_train,sample_weight=y_weight) pred = model.predict_proba(X_test) #model.fit(X_train.toarray(), y_train.toarray(),sample_weight=y_weight) #pred = model.predict_proba(X_test.toarray()) # model.fit(scipy.sparse.csr_matrix(X_train), scipy.sparse.csr_matrix(y_train),sample_weight=y_weight) # added tocsr() !!! # pred = model.predict_proba(scipy.sparse.csr_matrix(X_test)) #model.fit(scipy.sparse.csr_matrix(X_train), y_train,sample_weight=y_weight) #perceptron #pred = model.predict_proba(scipy.sparse.csr_matrix(X_test)) else: model.fit(X_train, y_train) pred = model.predict(X_test) if scale: if (y_train.shape[1] < 6): pred = np.apply_along_axis(lambda a: a/(np.max(a)-np.min(a)),1,pred) if is_normal and (y_train.shape[1] < 6): #pred[pred < 0.1] = 0.0 #pred[pred > 0.9] = 1.0 row_sums = pred.sum(axis=1) pred = pred / row_sums[:, np.newaxis] pred = np.around(pred,3) pred = pred.clip(0,1) if is_cv: return pred,y_true else: return pred def cv_loop(train, t2, model, is_sparse=True,kbest=0,is_class=False,is_tfidf=True,max_features=20000,perlabel=False,min_df=5,is_nominal=False,is_normal=False,is_LSA=False,scale=False): y = np.array(train.ix[:,4:]) ys = y[:,:5]#4:9 labeles of sentiment yw = y[:,5:9]#9:13 labels of when yk = y[:,9:]#13: labels of kind if is_class: ys,yw,yk = [np.around(y).astype(bool) for y in (ys,yw,yk)] if perlabel: pred,ytrue = vectorize(y,train,t2,model,is_tfidf = is_tfidf,kbest=kbest,is_sparse=is_sparse,max_features=max_features,is_cv=True,perlabel=perlabel,is_nominal=is_nominal,is_normal=is_normal,min_df=min_df,scale=scale) else: #(preds,ys_true),(predw,yw_true) = [vectorize(y,train,t2,model,is_tfidf = is_tfidf,kbest=kbest,is_sparse=is_sparse,max_features=max_features,is_cv=True,perlabel=perlabel,min_df=min_df,is_nominal=is_nominal,is_normal=is_normal) for y in (ys,yw)] #pred = np.hstack((preds,predw)) #ytrue = np.hstack((ys_true,yw_true)) (preds,ys_true),(predw,yw_true),(predk,yk_true) = [vectorize(y,train,t2,model,is_tfidf = is_tfidf,kbest=kbest,is_sparse=is_sparse,max_features=max_features,is_cv=True,perlabel=perlabel,min_df=min_df,is_nominal=is_nominal,is_normal=is_normal,is_LSA=is_LSA,scale=scale) for y in (ys,yw,yk)] pred = np.hstack((preds,predw,predk)) ytrue = np.hstack((ys_true,yw_true,yk_true)) #pred[pred < 0.01] = 0.0 #pred[pred > 0.99] = 1.0 mse = np.sqrt(np.sum(np.array(pred-ytrue)**2)/(ytrue.shape[0]*float(ytrue.shape[1]))) print 'Train error: {0}'.format(mse) return pred,ytrue def submission(predictions,filename='prediction.csv'): col = '%i,' + '%.2f,'*23 + '%.2f' header = "id,s1,s2,s3,s4,s5,w1,w2,w3,w4,k1,k2,k3,k4,k5,k6,k7,k8,k9,k10,k11,k12,k13,k14,k15" np.savetxt(filename, predictions,col, header=header,delimiter=',') # need to remove first two characters in the output file before submitting! def crossvalidate(clf,X,y,cv=3): scores=[] for i in range(int(y.shape[1])): clf.fit(X,y[:,i]) scores.append(cross_val_score(clf, X, y[:,i])) print scores[-1], return scores def predictMe(clf,X,y,test,ids): test_prediction=[] for i in range(int(y.shape[1])): clf.fit(X,y[:,i]) test_prediction.append(clf.predict(test)) testpred = np.array(test_prediction) prediction = np.array(np.hstack([np.matrix(ids).T, testpred.T])) return prediction def predictThis(clf,train,t2,kbest=0,max_features=20000,is_tfidf=True,is_sparse=True,is_nominal=False,is_LSA=False,min_df=5): y = np.array(train.ix[:,4:]) ys = y[:,:5]#4:9 labeles of sentiment yw = y[:,5:9]#9:13 labels of when yk = y[:,9:]#13: labels of kind if is_tfidf: #create a tf-idf class with the given features (stop_words='english' is removed since this is done in preprocessing) tfidf = TfidfVectorizer(max_features=max_features, strip_accents='unicode', analyzer='word', smooth_idf=True,sublinear_tf=True,max_df=0.5, min_df=min_df, ngram_range=(1,2)) sent, when, kind = [vectorize(y,train,t2,clf,is_tfidf = is_tfidf,kbest=kbest,is_sparse=is_sparse,max_features=max_features,is_nominal=is_nominal,is_LSA=is_LSA) for y in (ys,yw,yk)] testpred = np.hstack((sent, when, kind)) testpred = np.around(testpred.clip(0,1),3) prediction = np.array(np.hstack([np.matrix(t2['id']).T, testpred])) return prediction # to learn about indexing in pandas: http://pandas.pydata.org/pandas-docs/stable/indexing.html#advanced-indexing-with-hierarchical-index def predictKind(train_file,test_file): train_file='train.csv' test_file='test.csv' #read files into pandas train = p.read_csv(train_file) t2 = p.read_csv(test_file) for row in train.index: train['tweet'][row]=' '.join([train['tweet'][row],train['state'][row],str(train['location'][row])]) #preprocessing for kind prediction: emoticons and stop words can be ignored # for row in train.index: # train['tweet'][row] = pre.preprocess_pipeline(' '.join([train['tweet'][row],train['state'][row],str(train['location'][row])]), return_as_str=True, do_remove_stopwords=True,do_emoticons=True) # for row in t2.index: # t2['tweet'][row] = pre.preprocess_pipeline(' '.join([t2['tweet'][row],str(t2['state'][row]),str(t2['location'][row])]), return_as_str=True, do_remove_stopwords=True,do_emoticons=True) clf = linear_model.Ridge (alpha = 3.0, normalize = True) #pred,ytrue = cv_loop(train, t2, clf) #row_sums = pred.sum(axis=1) #pred_norm = pred / row_sums[:, numpy.newaxis] #mse = np.sqrt(np.sum(np.array(pred_norm-ytrue)**2)/(pred_norm.shape[0]*24.0)) #print 'Normalized train error: {0}'.format(mse) #Normalized train error: 0.366281924654 prediction = predictThis(clf,train,t2) submission(prediction,'prediction.csv') #metadata =sparse.csr_matrix([ metadata ]).T #X = sparse.hstack([X, metadata]).tocsr() #metadata = (metadata - mean(metadata))/(max(metadata) - min(metadata)) if __name__ == "__main__": # parse commandline arguments parser = argparse.ArgumentParser() parser.add_argument("train_file", help="a filename name must be specified") parser.add_argument("test_file", help="a filename name must be specified") args = parser.parse_args() prediction = predictKind(args.train_file, args.test_file) print 'elh' """ # a nice document classification example: http://scikit-learn.org/stable/auto_examples/document_classification_20newsgroups.html # we like ensemble: http://scikit-learn.org/stable/modules/ensemble.html You use the vocabulary parameter to specify what features should be used. For example, if you want only emoticons to be extracted, you can do the following: emoticons = {":)":0, ":P":1, ":(":2} vect = TfidfVectorizer(vocabulary=emoticons) matrix = vect.fit_transform(traindata) This will return a <Nx3 sparse matrix of type '<class 'numpy.int64'>' with M stored elements in Compressed Sparse Row format>]. Notice there are only 3 columns, one for each feature. If you want the vocabulary to include the emoticons as well as the N most common features, you could calculate the most frequent features first, then merge them with the emoticons and re-vectorize like so: # calculate the most frequent features first vect = TfidfVectorizer(vocabulary=emoticons) matrix = vect.fit_transform(traindata, max_features=10) top_features = vect.vocabulary_ n = len(top_features) # insert the emoticons into the vocabulary of common features emoticons = {":)":0, ":P":1, ":(":2)} for feature, index in emoticons.items(): top_features[feature] = n + index # re-vectorize using both sets of features # at this point len(top_features) == 13 vect = TfidfVectorizer(vocabulary=top_features) matrix = vect.fit_transform(traindata) """
mit
-7,789,753,431,757,353,000
42.592593
290
0.705091
false
sjdv1982/seamless
docs/plans/big-data/testing.py
1
12459
""" Performance tests based on tests/highlevel/high-in-low6-memory.py See also auth-*py. - Data overhead is now at ~6.5 ms / MB. A lot of the data overhead comes from json.dumps. This is to build mixed cells. The rest is from get_hash. Pure Python version (with get_hash and dumps) is at 6.3 ms / MB, so the rest of the data overhead is fine! - Database upload overhead is about the same (7 ms / MB) with a flatfile backend Database download is almost free. - A structured cell auth operation is about 10 ms. - map-list macro evaluation is cheap, 5 ms per (parallel) transformation - re-translation is about 20 ms per transformation (on top of the macro) - expression evaluation is about 10 ms + 0.5 ms / MB (of input + output) per transformation (speculative relationship!) - BUT: Non-linear scaling: between 100 and 1000 parallel transformations, a x4 slowdown is observed for the last three overheads above. NOTE: Mixed-to-str conversion is expensive, don't do it! """ import sys import seamless import seamless.core.execute seamless.core.execute.DIRECT_PRINT = True seamless.database_sink.connect() seamless.database_cache.connect() #seamless.set_ncores(2) #seamless.set_parallel_evaluations(5) seamless.set_ncores(8) ### seamless.set_parallel_evaluations(100) ### # for the big testing, 20 evaluations seamless.set_parallel_evaluations(20) ### """ import logging logging.basicConfig() logging.getLogger("seamless").setLevel(logging.DEBUG) """ from seamless.highlevel import Context, Cell, Macro from seamless.highlevel.library import LibraryContainer import time import cProfile cProfile.profiler = cProfile.Profile() mylib = LibraryContainer("mylib") mylib.map_list_N = Context() def constructor(ctx, libctx, context_graph, inp, result): m = ctx.m = Macro() m.graph = context_graph m.pins.result = {"io": "output", "celltype": "mixed", "hash_pattern": {"!": "#"}} ctx.inp = Context() ctx.cs_inp = Context() inp_prefix = "INPUT_" m.inp_prefix = inp_prefix for key in inp: c = Cell() ctx.inp[key] = c c.hash_pattern = {"!": "#"} inp[key].connect(c) ctx.cs_inp[key] = Cell("checksum") ctx.cs_inp[key] = ctx.inp[key] setattr(m, inp_prefix + key , ctx.cs_inp[key]) def map_list_N(ctx, inp_prefix, graph, **inp): #print("INP", inp) first_k = list(inp.keys())[0] length = len(inp[first_k]) first_k = first_k[len(inp_prefix):] for k0 in inp: k = k0[len(inp_prefix):] if len(inp[k0]) != length: err = "all cells in inp must have the same length, but '{}' has length {} while '{}' has length {}" raise ValueError(err.format(k, len(inp[k0]), first_k, length)) print("LENGTH", length) from seamless.core import Cell as CoreCell from seamless.core.unbound_context import UnboundContext pseudo_connections = [] ctx.result = cell("mixed", hash_pattern = {"!": "#"}) ctx.sc_data = cell("mixed", hash_pattern = {"!": "#"}) ctx.sc_buffer = cell("mixed", hash_pattern = {"!": "#"}) ctx.sc = StructuredCell( data=ctx.sc_data, buffer=ctx.sc_buffer, inchannels=[(n,) for n in range(length)], outchannels=[()], hash_pattern = {"!": "#"} ) for n in range(length): #print("MACRO", n+1) hc = HighLevelContext(graph) subctx = "subctx%d" % (n+1) setattr(ctx, subctx, hc) if not hasattr(hc, "inp"): raise TypeError("map_list_N context must have a subcontext called 'inp'") hci = hc.inp if not isinstance(hci, UnboundContext): raise TypeError("map_list_N context must have an attribute 'inp' that is a context, not a {}".format(type(hci))) for k0 in inp: k = k0[len(inp_prefix):] if not hasattr(hci, k): raise TypeError("map_list_N context must have a cell called inp.'{}'".format(k)) if isinstance(hci[k], StructuredCell): raise TypeError("map_list_N context has a cell called inp.'{}', but its celltype must be mixed, not structured".format(k)) if not isinstance(hci[k], CoreCell): raise TypeError("map_list_N context must have an attribute inp.'{}' that is a cell, not a {}".format(k, type(hci[k]))) if hci[k].celltype != "mixed": raise TypeError("map_list_N context has a cell called inp.'{}', but its celltype must be mixed, not {}".format(k, hci[k].celltype)) con = [".." + k], ["ctx", subctx, "inp", k] pseudo_connections.append(con) cs = inp[k0][n] hci[k].set_checksum(cs) resultname = "result%d" % (n+1) setattr(ctx, resultname, cell("mixed")) c = getattr(ctx, resultname) hc.result.connect(c) c.connect(ctx.sc.inchannels[(n,)]) con = ["ctx", subctx, "result"], ["..result"] pseudo_connections.append(con) ctx.sc.outchannels[()].connect(ctx.result) ctx._pseudo_connections = pseudo_connections print("/MACRO") """ import logging logging.basicConfig() logging.getLogger("seamless").setLevel(logging.DEBUG) """ import cProfile cProfile.profiler.enable() m.code = map_list_N ctx.result = Cell() ctx.result.hash_pattern = {"!": "#"} ctx.result = m.result result.connect_from(ctx.result) mylib.map_list_N.constructor = constructor mylib.map_list_N.params = { "context_graph": "context", "inp": { "type": "celldict", "io": "input" }, "result": { "type": "cell", "io": "output" }, } ctx = Context() ctx.adder = Context() sctx = ctx.adder sctx.inp = Context() sctx.inp.a = Cell("mixed") sctx.inp.b = Cell("mixed") sctx.a = Cell("mixed") sctx.b = Cell("mixed") sctx.a = sctx.inp.a sctx.b = sctx.inp.b def add(a,b): print("ADD", a[:10]) return a+b sctx.add = add sctx.add.a = sctx.a sctx.add.b = sctx.b sctx.result = sctx.add sctx.result.celltype = "mixed" ctx.compute() ctx.data_a = Cell() ctx.data_a.hash_pattern = {"!": "#"} ctx.data_b = Cell() ctx.data_b.hash_pattern = {"!": "#"} ctx.compute() # Next section is 14.5 secs (if the database is filled), but can be elided to ~0.5s by setting checksum directly (if in flatfile cache). # Not having a DB at all is also 13 secs, so DB request communication (without upload) doesn't cost much. repeat = int(10e6) #repeat = int(5) #for n in range(1000): # 2x10 GB #for n in range(100): # 2x1 GB for n in range(1000): a = "A:%d:" % n + str(n%10) * repeat b = "B:%d:" % n + str(n%10) * repeat ctx.data_a[n] = a ctx.data_b[n] = b if n % 20 == 0: ctx.compute() print(n+1) ctx.compute() print(ctx.data_a.checksum) print(ctx.data_b.checksum) """ ctx.data_a.set_checksum("d07050610c50de8c810aa1d1e322786ed8932cf6eafa0fbe1f132b2c881af9c2") ctx.data_b.set_checksum("374c02504f89ed0a760b03c3e1fd2258988576b919d763254709b66fc7bfb22e") ctx.compute() """ # ### For repeat=10 million, 1000 items ### ctx.data_a.set_checksum("fa4e6aa7e7edaa6feb036fd5e8c28ffc48575cefc332187552c5be4bf0511af8") ### ctx.data_b.set_checksum("2988c44780790e4ffceb1f97391e475f165e316f27a656957282a2998aee9d4f") ### For repeat=10 million, 200 items ### ctx.data_a.set_checksum("d07050610c50de8c810aa1d1e322786ed8932cf6eafa0fbe1f132b2c881af9c2") ### ctx.data_b.set_checksum("374c02504f89ed0a760b03c3e1fd2258988576b919d763254709b66fc7bfb22e") ### For repeat=10 million ### ctx.data_a.set_checksum("983730afb7ab41d524b72f1097daaf4a3c15b98943291f96e523730849cabe8c") ### ctx.data_b.set_checksum("46dabc02b59be44064a9e06dd50bc6841833578c2b6339fbc43f090cc17831fa") ### For repeat=5 ### ctx.data_a.set_checksum("9b4a551a6c1c5830d6070b9c22ae1788b9743e9637be47d56103bcda019a897c") ### ctx.data_b.set_checksum("9820f1ab795db7b0d195f21966ecb071d76e9ce2fd3a90845974a3905584eb3e") ctx.compute() """ If there is no database (100 x repeat 10e6): - 13 secs up to here (6.5 ms per MB) - 0.5 secs to evaluate the macro - 2.3 secs (2.8 - 0.5) for re-translation (23 ms per transformer) - 32 secs total time, which leaves 32 - 13 - 0.5 = 18.5 secs for transformation and expression evaluation Since 13 secs is required for calculate checksum and decoding, that means ~5.5 secs (55 ms per transformer) overhead This is a supplement of 32 ms over just re-translation If there is no database (100 x repeat 5): - 2.3 secs up to here (12 ms per auth operation) - Still 0.5 secs to evaluate the macro - Still 2.3 secs (2.8 - 0.5) for re-translation (23 ms per transformer, independent of data size!) - 6.2 secs total time, which leaves 6.2 - 2.3 - 0.5 = 3.5 secs for transformation and expression evaluation This is an overhead of 35 ms per transformer, a supplement of just 12 ms over re-translation The 20 ms reduction compared to above comes from not handling 2x10 MB of input and 20 MB of output, so that's 0.5 ms/MB. If there is no database (1000 x repeat 5): - 11.7 secs up to here (12 ms per auth operation). So scales linearly. - 6.5 secs to evaluate the macro, so scales about linearly - 98 secs (104.5 - 6.5) for re-translation, which is 4x slower than above (98 ms) - 145 secs total time, which leaves 145 - 11.7 - 6.5 = 127 secs for transformation and expression evaluation This is an overhead of 127 ms per transformer, which is 4x slower than above (127 ms). => So in principle, 90 sconds slower than might be - Some 45 secs is await-upon-connection-tasks, this could be optimized? - 12 seconds from isinstance is probably unavoidable - 9 seconds comes from validate deep structure, that may be unavoidable - 5 seconds each from taskmanager.add_task (61k tasks) and asyncio.Task.done (119 million tasks). Avoidable? => do maplist-inside-maplist If the database has not been filled: - 27.5 secs up to here If the database has been filled: - 14 secs up to here: to synthesize the data, and to verify that all is present So the raw upload is 13.5 seconds (27.5 - 14); and communication with the DB delays only 1 sec. - 1.5 secs up to here, with the above elision. With the database: - 1.5 secs to evaluate the macro (DB slows down!) - 5.5 secs for re-translation - 45.7 secs total time, which leaves 45.7 - 5.5 - 1.5 = 38.5 secs for transformation and expression evaluation Compare this to the 18.5 secs w/o database, this is a loss of 20 secs. But we have to count the download of the inputs and upload of the results. When removing the tfr entries from the database, transformations will be repeated, but no buffers will be uploaded, as the sink knows them already. This brings total time down to 32 secs, the same as no database! So all of the extra overhead is from upload, and download is almost free. (This could be hard disk caching, though) - 5.5 secs total time with pulling transformation results out of the DB. Again, download is almost free. Big test with the database (1000 x repeat 10e6): - Total time 940 secs. Data upload overhead should be ~120 secs, and Seamless data overhead should be ~140 secs. - 142 secs for re-translation + macro evaluation (142 ms / transformation), a factor 6 slowdown - 940 - 142 - 120 - 140 = ~540 secs for evaluation I.e. 540 ms per transformation. If the same slowdown applies, it would have been 90. But we would have expected 30. So a larger slowdown (fewer parallel expressions may have been a cause too) """ ctx.result = Cell() ctx.result.hash_pattern = {"!": "#"} ctx.compute() ctx.include(mylib.map_list_N) ctx.inst = ctx.lib.map_list_N( context_graph = ctx.adder, inp = {"a": ctx.data_a, "b": ctx.data_b}, result = ctx.result ) ctx.compute() print("Exception:", ctx.inst.ctx.m.exception) print(ctx.result.data) import pstats sortby = 'tottime' ps = pstats.Stats(cProfile.profiler).sort_stats(sortby) ps.print_stats(40) t0 = time.time() """ print("Re-set") graph = ctx.get_graph() ctx_dummy = Context() dummy_graph = ctx_dummy.get_graph() ctx.set_graph(dummy_graph) ctx.translate(force=True) ctx.compute() print(time.time()-t0) print("Re-eval") ctx.set_graph(graph) """ """ ctx.translate(force=True) ctx.compute() print(time.time()-t0) """
mit
-8,654,298,033,990,642,000
34.804598
151
0.662332
false
PuzzleboxIO/jigsaw-python
Puzzlebox/Jigsaw/Design_Plugin_Eeg.py
1
34778
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'interface/puzzlebox_jigsaw_interface_design-plugin_eeg.ui' # # Created: Sun May 10 18:28:33 2015 # by: pyside-uic 0.2.15 running on PySide 1.2.2 # # WARNING! All changes made in this file will be lost! from PySide import QtCore, QtGui class Ui_Form(object): def setupUi(self, Form): Form.setObjectName("Form") #Form.resize() Form.setAutoFillBackground(False) self.verticalLayoutWidget_2 = QtGui.QWidget(Form) self.verticalLayoutWidget_2.setGeometry(QtCore.QRect(0, 0, 975, 776)) self.verticalLayoutWidget_2.setObjectName("verticalLayoutWidget_2") self.verticalLayout = QtGui.QVBoxLayout(self.verticalLayoutWidget_2) self.verticalLayout.setContentsMargins(0, 0, 0, 0) self.verticalLayout.setObjectName("verticalLayout") self.horizontalLayout_7 = QtGui.QHBoxLayout() self.horizontalLayout_7.setContentsMargins(4, 4, 4, 4) self.horizontalLayout_7.setObjectName("horizontalLayout_7") spacerItem = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayout_7.addItem(spacerItem) self.verticalLayoutEEGSource = QtGui.QVBoxLayout() self.verticalLayoutEEGSource.setSpacing(6) self.verticalLayoutEEGSource.setObjectName("verticalLayoutEEGSource") self.labelEEGHeadsetModel = QtGui.QLabel(self.verticalLayoutWidget_2) self.labelEEGHeadsetModel.setAlignment(QtCore.Qt.AlignCenter) self.labelEEGHeadsetModel.setObjectName("labelEEGHeadsetModel") self.verticalLayoutEEGSource.addWidget(self.labelEEGHeadsetModel) self.comboBoxEEGHeadsetModel = QtGui.QComboBox(self.verticalLayoutWidget_2) self.comboBoxEEGHeadsetModel.setEnabled(True) self.comboBoxEEGHeadsetModel.setFocusPolicy(QtCore.Qt.NoFocus) self.comboBoxEEGHeadsetModel.setObjectName("comboBoxEEGHeadsetModel") self.comboBoxEEGHeadsetModel.addItem("") self.comboBoxEEGHeadsetModel.addItem("") self.comboBoxEEGHeadsetModel.addItem("") self.comboBoxEEGHeadsetModel.addItem("") self.comboBoxEEGHeadsetModel.addItem("") self.verticalLayoutEEGSource.addWidget(self.comboBoxEEGHeadsetModel) self.line_3 = QtGui.QFrame(self.verticalLayoutWidget_2) self.line_3.setFrameShape(QtGui.QFrame.HLine) self.line_3.setFrameShadow(QtGui.QFrame.Sunken) self.line_3.setObjectName("line_3") self.verticalLayoutEEGSource.addWidget(self.line_3) self.labelEEGServerSource = QtGui.QLabel(self.verticalLayoutWidget_2) self.labelEEGServerSource.setAlignment(QtCore.Qt.AlignCenter) self.labelEEGServerSource.setObjectName("labelEEGServerSource") self.verticalLayoutEEGSource.addWidget(self.labelEEGServerSource) self.comboBoxEEGSource = QtGui.QComboBox(self.verticalLayoutWidget_2) self.comboBoxEEGSource.setEnabled(True) self.comboBoxEEGSource.setObjectName("comboBoxEEGSource") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.comboBoxEEGSource.addItem("") self.verticalLayoutEEGSource.addWidget(self.comboBoxEEGSource) self.comboBoxDeviceSelect = QtGui.QComboBox(self.verticalLayoutWidget_2) self.comboBoxDeviceSelect.setObjectName("comboBoxDeviceSelect") self.comboBoxDeviceSelect.addItem("") self.verticalLayoutEEGSource.addWidget(self.comboBoxDeviceSelect) self.pushButtonControlSearch = QtGui.QPushButton(self.verticalLayoutWidget_2) self.pushButtonControlSearch.setObjectName("pushButtonControlSearch") self.verticalLayoutEEGSource.addWidget(self.pushButtonControlSearch) self.checkBoxControlEmulateThinkGear = QtGui.QCheckBox(self.verticalLayoutWidget_2) self.checkBoxControlEmulateThinkGear.setChecked(True) self.checkBoxControlEmulateThinkGear.setObjectName("checkBoxControlEmulateThinkGear") self.verticalLayoutEEGSource.addWidget(self.checkBoxControlEmulateThinkGear) self.lineControlSourceServer = QtGui.QFrame(self.verticalLayoutWidget_2) self.lineControlSourceServer.setFrameShape(QtGui.QFrame.HLine) self.lineControlSourceServer.setFrameShadow(QtGui.QFrame.Sunken) self.lineControlSourceServer.setObjectName("lineControlSourceServer") self.verticalLayoutEEGSource.addWidget(self.lineControlSourceServer) self.checkBoxControlEnableServer = QtGui.QCheckBox(self.verticalLayoutWidget_2) self.checkBoxControlEnableServer.setChecked(True) self.checkBoxControlEnableServer.setObjectName("checkBoxControlEnableServer") self.verticalLayoutEEGSource.addWidget(self.checkBoxControlEnableServer) self.formLayoutControlHostPort = QtGui.QFormLayout() self.formLayoutControlHostPort.setSizeConstraint(QtGui.QLayout.SetMinimumSize) self.formLayoutControlHostPort.setObjectName("formLayoutControlHostPort") self.textLabelSynapseHost = QtGui.QLabel(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.textLabelSynapseHost.sizePolicy().hasHeightForWidth()) self.textLabelSynapseHost.setSizePolicy(sizePolicy) self.textLabelSynapseHost.setAlignment(QtCore.Qt.AlignCenter) self.textLabelSynapseHost.setObjectName("textLabelSynapseHost") self.formLayoutControlHostPort.setWidget(0, QtGui.QFormLayout.LabelRole, self.textLabelSynapseHost) self.lineEditSynapseHost = QtGui.QLineEdit(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Fixed) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.lineEditSynapseHost.sizePolicy().hasHeightForWidth()) self.lineEditSynapseHost.setSizePolicy(sizePolicy) self.lineEditSynapseHost.setAlignment(QtCore.Qt.AlignHCenter) self.lineEditSynapseHost.setObjectName("lineEditSynapseHost") self.formLayoutControlHostPort.setWidget(0, QtGui.QFormLayout.FieldRole, self.lineEditSynapseHost) self.textLabelSynapsePort = QtGui.QLabel(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.textLabelSynapsePort.sizePolicy().hasHeightForWidth()) self.textLabelSynapsePort.setSizePolicy(sizePolicy) self.textLabelSynapsePort.setAlignment(QtCore.Qt.AlignCenter) self.textLabelSynapsePort.setWordWrap(False) self.textLabelSynapsePort.setObjectName("textLabelSynapsePort") self.formLayoutControlHostPort.setWidget(1, QtGui.QFormLayout.LabelRole, self.textLabelSynapsePort) self.lineEditSynapsePort = QtGui.QLineEdit(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Fixed) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.lineEditSynapsePort.sizePolicy().hasHeightForWidth()) self.lineEditSynapsePort.setSizePolicy(sizePolicy) self.lineEditSynapsePort.setAlignment(QtCore.Qt.AlignHCenter) self.lineEditSynapsePort.setObjectName("lineEditSynapsePort") self.formLayoutControlHostPort.setWidget(1, QtGui.QFormLayout.FieldRole, self.lineEditSynapsePort) self.verticalLayoutEEGSource.addLayout(self.formLayoutControlHostPort) self.horizontalLayout_3 = QtGui.QHBoxLayout() self.horizontalLayout_3.setObjectName("horizontalLayout_3") self.pushButtonSynapseServer = QtGui.QPushButton(self.verticalLayoutWidget_2) self.pushButtonSynapseServer.setMinimumSize(QtCore.QSize(0, 64)) self.pushButtonSynapseServer.setObjectName("pushButtonSynapseServer") self.horizontalLayout_3.addWidget(self.pushButtonSynapseServer) self.verticalLayoutEEGSource.addLayout(self.horizontalLayout_3) spacerItem1 = QtGui.QSpacerItem(20, 40, QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Expanding) self.verticalLayoutEEGSource.addItem(spacerItem1) self.horizontalLayout_7.addLayout(self.verticalLayoutEEGSource) self.verticalLayoutEEGPlot = QtGui.QVBoxLayout() self.verticalLayoutEEGPlot.setSizeConstraint(QtGui.QLayout.SetMinimumSize) self.verticalLayoutEEGPlot.setObjectName("verticalLayoutEEGPlot") self.horizontalLayoutChartTop = QtGui.QHBoxLayout() self.horizontalLayoutChartTop.setObjectName("horizontalLayoutChartTop") spacerItem2 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayoutChartTop.addItem(spacerItem2) self.labelChartTop = QtGui.QLabel(self.verticalLayoutWidget_2) self.labelChartTop.setObjectName("labelChartTop") self.horizontalLayoutChartTop.addWidget(self.labelChartTop) spacerItem3 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayoutChartTop.addItem(spacerItem3) self.verticalLayoutEEGPlot.addLayout(self.horizontalLayoutChartTop) self.widgetPlotRawEEG = QtGui.QWidget(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.widgetPlotRawEEG.sizePolicy().hasHeightForWidth()) self.widgetPlotRawEEG.setSizePolicy(sizePolicy) self.widgetPlotRawEEG.setMinimumSize(QtCore.QSize(520, 160)) self.widgetPlotRawEEG.setObjectName("widgetPlotRawEEG") self.labelMockupRawEEG = QtGui.QLabel(self.widgetPlotRawEEG) self.labelMockupRawEEG.setGeometry(QtCore.QRect(0, 0, 520, 160)) self.labelMockupRawEEG.setText("") self.labelMockupRawEEG.setPixmap(QtGui.QPixmap("../../../../.designer/backup/images/mockup-raw_eeg_waves.png")) self.labelMockupRawEEG.setObjectName("labelMockupRawEEG") self.verticalLayoutEEGPlot.addWidget(self.widgetPlotRawEEG) self.widgetPlotHistoryEmotivCognitiv = QtGui.QWidget(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.widgetPlotHistoryEmotivCognitiv.sizePolicy().hasHeightForWidth()) self.widgetPlotHistoryEmotivCognitiv.setSizePolicy(sizePolicy) self.widgetPlotHistoryEmotivCognitiv.setMinimumSize(QtCore.QSize(520, 160)) self.widgetPlotHistoryEmotivCognitiv.setObjectName("widgetPlotHistoryEmotivCognitiv") self.labelMockupHistoryEmotivCognitiv = QtGui.QLabel(self.widgetPlotHistoryEmotivCognitiv) self.labelMockupHistoryEmotivCognitiv.setGeometry(QtCore.QRect(0, 0, 520, 160)) self.labelMockupHistoryEmotivCognitiv.setText("") self.labelMockupHistoryEmotivCognitiv.setPixmap(QtGui.QPixmap("../../../../.designer/backup/images/mockup-esense_values_history.jpg")) self.labelMockupHistoryEmotivCognitiv.setScaledContents(True) self.labelMockupHistoryEmotivCognitiv.setObjectName("labelMockupHistoryEmotivCognitiv") self.verticalLayoutEEGPlot.addWidget(self.widgetPlotHistoryEmotivCognitiv) self.horizontalLayoutChartBottom = QtGui.QHBoxLayout() self.horizontalLayoutChartBottom.setObjectName("horizontalLayoutChartBottom") spacerItem4 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayoutChartBottom.addItem(spacerItem4) self.labelChartBottom = QtGui.QLabel(self.verticalLayoutWidget_2) self.labelChartBottom.setTextFormat(QtCore.Qt.AutoText) self.labelChartBottom.setObjectName("labelChartBottom") self.horizontalLayoutChartBottom.addWidget(self.labelChartBottom) spacerItem5 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayoutChartBottom.addItem(spacerItem5) self.verticalLayoutEEGPlot.addLayout(self.horizontalLayoutChartBottom) self.widgetPlotHistoryThinkGear = QtGui.QWidget(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.widgetPlotHistoryThinkGear.sizePolicy().hasHeightForWidth()) self.widgetPlotHistoryThinkGear.setSizePolicy(sizePolicy) self.widgetPlotHistoryThinkGear.setMinimumSize(QtCore.QSize(520, 160)) self.widgetPlotHistoryThinkGear.setObjectName("widgetPlotHistoryThinkGear") self.labelMockupHistoryThinkGear = QtGui.QLabel(self.widgetPlotHistoryThinkGear) self.labelMockupHistoryThinkGear.setGeometry(QtCore.QRect(0, 0, 520, 160)) self.labelMockupHistoryThinkGear.setText("") self.labelMockupHistoryThinkGear.setPixmap(QtGui.QPixmap("../../../../.designer/backup/images/mockup-esense_values_history.jpg")) self.labelMockupHistoryThinkGear.setScaledContents(True) self.labelMockupHistoryThinkGear.setObjectName("labelMockupHistoryThinkGear") self.verticalLayoutEEGPlot.addWidget(self.widgetPlotHistoryThinkGear) self.widgetPlotHistoryEmotivAffectiv = QtGui.QWidget(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.widgetPlotHistoryEmotivAffectiv.sizePolicy().hasHeightForWidth()) self.widgetPlotHistoryEmotivAffectiv.setSizePolicy(sizePolicy) self.widgetPlotHistoryEmotivAffectiv.setMinimumSize(QtCore.QSize(520, 160)) self.widgetPlotHistoryEmotivAffectiv.setObjectName("widgetPlotHistoryEmotivAffectiv") self.labelMockupHistoryEmotivAffectiv = QtGui.QLabel(self.widgetPlotHistoryEmotivAffectiv) self.labelMockupHistoryEmotivAffectiv.setGeometry(QtCore.QRect(0, 0, 520, 160)) self.labelMockupHistoryEmotivAffectiv.setText("") self.labelMockupHistoryEmotivAffectiv.setPixmap(QtGui.QPixmap("../../../../.designer/backup/images/mockup-esense_values_history.jpg")) self.labelMockupHistoryEmotivAffectiv.setScaledContents(True) self.labelMockupHistoryEmotivAffectiv.setObjectName("labelMockupHistoryEmotivAffectiv") self.verticalLayoutEEGPlot.addWidget(self.widgetPlotHistoryEmotivAffectiv) spacerItem6 = QtGui.QSpacerItem(20, 40, QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Expanding) self.verticalLayoutEEGPlot.addItem(spacerItem6) self.horizontalLayout_7.addLayout(self.verticalLayoutEEGPlot) self.verticalLayoutSessionProfile = QtGui.QVBoxLayout() self.verticalLayoutSessionProfile.setObjectName("verticalLayoutSessionProfile") self.label = QtGui.QLabel(self.verticalLayoutWidget_2) self.label.setObjectName("label") self.verticalLayoutSessionProfile.addWidget(self.label) self.horizontalLayout_4 = QtGui.QHBoxLayout() self.horizontalLayout_4.setSizeConstraint(QtGui.QLayout.SetMinimumSize) self.horizontalLayout_4.setObjectName("horizontalLayout_4") self.textLabelControlConcentration = QtGui.QLabel(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.textLabelControlConcentration.sizePolicy().hasHeightForWidth()) self.textLabelControlConcentration.setSizePolicy(sizePolicy) self.textLabelControlConcentration.setWordWrap(False) self.textLabelControlConcentration.setObjectName("textLabelControlConcentration") self.horizontalLayout_4.addWidget(self.textLabelControlConcentration) spacerItem7 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayout_4.addItem(spacerItem7) self.pushButtonControlConcentrationEnable = QtGui.QPushButton(self.verticalLayoutWidget_2) self.pushButtonControlConcentrationEnable.setEnabled(False) self.pushButtonControlConcentrationEnable.setCheckable(True) self.pushButtonControlConcentrationEnable.setChecked(True) self.pushButtonControlConcentrationEnable.setObjectName("pushButtonControlConcentrationEnable") self.horizontalLayout_4.addWidget(self.pushButtonControlConcentrationEnable) self.verticalLayoutSessionProfile.addLayout(self.horizontalLayout_4) self.progressBarControlConcentration = QtGui.QProgressBar(self.verticalLayoutWidget_2) self.progressBarControlConcentration.setEnabled(True) self.progressBarControlConcentration.setMinimumSize(QtCore.QSize(0, 60)) self.progressBarControlConcentration.setProperty("value", 0) self.progressBarControlConcentration.setOrientation(QtCore.Qt.Horizontal) self.progressBarControlConcentration.setObjectName("progressBarControlConcentration") self.verticalLayoutSessionProfile.addWidget(self.progressBarControlConcentration) self.horizontalLayout_5 = QtGui.QHBoxLayout() self.horizontalLayout_5.setSizeConstraint(QtGui.QLayout.SetMinimumSize) self.horizontalLayout_5.setObjectName("horizontalLayout_5") self.textLabelControlRelaxation = QtGui.QLabel(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.textLabelControlRelaxation.sizePolicy().hasHeightForWidth()) self.textLabelControlRelaxation.setSizePolicy(sizePolicy) self.textLabelControlRelaxation.setWordWrap(False) self.textLabelControlRelaxation.setObjectName("textLabelControlRelaxation") self.horizontalLayout_5.addWidget(self.textLabelControlRelaxation) spacerItem8 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayout_5.addItem(spacerItem8) self.pushButtonControlRelaxationEnable = QtGui.QPushButton(self.verticalLayoutWidget_2) self.pushButtonControlRelaxationEnable.setEnabled(False) self.pushButtonControlRelaxationEnable.setCheckable(True) self.pushButtonControlRelaxationEnable.setChecked(True) self.pushButtonControlRelaxationEnable.setObjectName("pushButtonControlRelaxationEnable") self.horizontalLayout_5.addWidget(self.pushButtonControlRelaxationEnable) self.verticalLayoutSessionProfile.addLayout(self.horizontalLayout_5) self.progressBarControlRelaxation = QtGui.QProgressBar(self.verticalLayoutWidget_2) self.progressBarControlRelaxation.setEnabled(True) self.progressBarControlRelaxation.setMinimumSize(QtCore.QSize(0, 60)) self.progressBarControlRelaxation.setProperty("value", 0) self.progressBarControlRelaxation.setOrientation(QtCore.Qt.Horizontal) self.progressBarControlRelaxation.setObjectName("progressBarControlRelaxation") self.verticalLayoutSessionProfile.addWidget(self.progressBarControlRelaxation) spacerItem9 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.verticalLayoutSessionProfile.addItem(spacerItem9) self.horizontalLayout_6 = QtGui.QHBoxLayout() self.horizontalLayout_6.setObjectName("horizontalLayout_6") self.textLabelControlConnectionLevel = QtGui.QLabel(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.textLabelControlConnectionLevel.sizePolicy().hasHeightForWidth()) self.textLabelControlConnectionLevel.setSizePolicy(sizePolicy) self.textLabelControlConnectionLevel.setWordWrap(False) self.textLabelControlConnectionLevel.setObjectName("textLabelControlConnectionLevel") self.horizontalLayout_6.addWidget(self.textLabelControlConnectionLevel) spacerItem10 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayout_6.addItem(spacerItem10) self.verticalLayoutSessionProfile.addLayout(self.horizontalLayout_6) self.progressBarControlConnectionLevel = QtGui.QProgressBar(self.verticalLayoutWidget_2) self.progressBarControlConnectionLevel.setEnabled(True) self.progressBarControlConnectionLevel.setProperty("value", 0) self.progressBarControlConnectionLevel.setOrientation(QtCore.Qt.Horizontal) self.progressBarControlConnectionLevel.setObjectName("progressBarControlConnectionLevel") self.verticalLayoutSessionProfile.addWidget(self.progressBarControlConnectionLevel) self.line_6 = QtGui.QFrame(self.verticalLayoutWidget_2) self.line_6.setFrameShape(QtGui.QFrame.HLine) self.line_6.setFrameShadow(QtGui.QFrame.Sunken) self.line_6.setObjectName("line_6") self.verticalLayoutSessionProfile.addWidget(self.line_6) self.formLayoutSessionData = QtGui.QFormLayout() self.formLayoutSessionData.setFieldGrowthPolicy(QtGui.QFormLayout.AllNonFixedFieldsGrow) self.formLayoutSessionData.setObjectName("formLayoutSessionData") self.textLabelSessionTimeTitle = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelSessionTimeTitle.setAlignment(QtCore.Qt.AlignCenter) self.textLabelSessionTimeTitle.setObjectName("textLabelSessionTimeTitle") self.formLayoutSessionData.setWidget(0, QtGui.QFormLayout.LabelRole, self.textLabelSessionTimeTitle) self.textLabelSessionTime = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelSessionTime.setAlignment(QtCore.Qt.AlignCenter) self.textLabelSessionTime.setObjectName("textLabelSessionTime") self.formLayoutSessionData.setWidget(0, QtGui.QFormLayout.FieldRole, self.textLabelSessionTime) self.textLabelPacketsReceivedTitle = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelPacketsReceivedTitle.setAlignment(QtCore.Qt.AlignCenter) self.textLabelPacketsReceivedTitle.setObjectName("textLabelPacketsReceivedTitle") self.formLayoutSessionData.setWidget(1, QtGui.QFormLayout.LabelRole, self.textLabelPacketsReceivedTitle) self.textLabelPacketsReceived = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelPacketsReceived.setAlignment(QtCore.Qt.AlignCenter) self.textLabelPacketsReceived.setObjectName("textLabelPacketsReceived") self.formLayoutSessionData.setWidget(1, QtGui.QFormLayout.FieldRole, self.textLabelPacketsReceived) self.textLabelPacketsDroppedTitle = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelPacketsDroppedTitle.setAlignment(QtCore.Qt.AlignCenter) self.textLabelPacketsDroppedTitle.setObjectName("textLabelPacketsDroppedTitle") self.formLayoutSessionData.setWidget(2, QtGui.QFormLayout.LabelRole, self.textLabelPacketsDroppedTitle) self.textLabelPacketsDropped = QtGui.QLabel(self.verticalLayoutWidget_2) self.textLabelPacketsDropped.setAlignment(QtCore.Qt.AlignCenter) self.textLabelPacketsDropped.setObjectName("textLabelPacketsDropped") self.formLayoutSessionData.setWidget(2, QtGui.QFormLayout.FieldRole, self.textLabelPacketsDropped) self.verticalLayoutSessionProfile.addLayout(self.formLayoutSessionData) spacerItem11 = QtGui.QSpacerItem(20, 40, QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Expanding) self.verticalLayoutSessionProfile.addItem(spacerItem11) self.horizontalLayout_7.addLayout(self.verticalLayoutSessionProfile) spacerItem12 = QtGui.QSpacerItem(40, 20, QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Minimum) self.horizontalLayout_7.addItem(spacerItem12) self.verticalLayout.addLayout(self.horizontalLayout_7) self.webViewEEG = QtWebKit.QWebView(self.verticalLayoutWidget_2) sizePolicy = QtGui.QSizePolicy(QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Expanding) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.webViewEEG.sizePolicy().hasHeightForWidth()) self.webViewEEG.setSizePolicy(sizePolicy) self.webViewEEG.setObjectName("webViewEEG") self.verticalLayout.addWidget(self.webViewEEG) spacerItem13 = QtGui.QSpacerItem(20, 40, QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Expanding) self.verticalLayout.addItem(spacerItem13) self.retranslateUi(Form) QtCore.QMetaObject.connectSlotsByName(Form) def retranslateUi(self, Form): Form.setWindowTitle(QtGui.QApplication.translate("Form", "Puzzlebox Jigsaw - Plug-in - EEG", None, QtGui.QApplication.UnicodeUTF8)) self.labelEEGHeadsetModel.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">EEG Headset Model</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGHeadsetModel.setItemText(0, QtGui.QApplication.translate("Form", "NeuroSky MindWave Mobile", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGHeadsetModel.setItemText(1, QtGui.QApplication.translate("Form", "NeuroSky MindWave", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGHeadsetModel.setItemText(2, QtGui.QApplication.translate("Form", "NeuroSky MindSet", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGHeadsetModel.setItemText(3, QtGui.QApplication.translate("Form", "Emotiv EPOC", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGHeadsetModel.setItemText(4, QtGui.QApplication.translate("Form", "InterAxon Muse", None, QtGui.QApplication.UnicodeUTF8)) self.labelEEGServerSource.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">EEG Data Source</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(0, QtGui.QApplication.translate("Form", "Hardware Device", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(1, QtGui.QApplication.translate("Form", "ThinkGear Emulator", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(2, QtGui.QApplication.translate("Form", "Puzzlebox Synapse", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(3, QtGui.QApplication.translate("Form", "ThinkGear Connect", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(4, QtGui.QApplication.translate("Form", "BCI2000", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(5, QtGui.QApplication.translate("Form", "MuseIO", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxEEGSource.setItemText(6, QtGui.QApplication.translate("Form", "Spacebrew", None, QtGui.QApplication.UnicodeUTF8)) self.comboBoxDeviceSelect.setItemText(0, QtGui.QApplication.translate("Form", "No Devices Found", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonControlSearch.setText(QtGui.QApplication.translate("Form", "Search", None, QtGui.QApplication.UnicodeUTF8)) self.checkBoxControlEmulateThinkGear.setText(QtGui.QApplication.translate("Form", "Emulate NeuroSky ThinkGear", None, QtGui.QApplication.UnicodeUTF8)) self.checkBoxControlEnableServer.setText(QtGui.QApplication.translate("Form", "Enable Synapse Server", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelSynapseHost.setText(QtGui.QApplication.translate("Form", "Host", None, QtGui.QApplication.UnicodeUTF8)) self.lineEditSynapseHost.setText(QtGui.QApplication.translate("Form", "N/A", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelSynapsePort.setText(QtGui.QApplication.translate("Form", "<p align=\"center\">Port</p>", None, QtGui.QApplication.UnicodeUTF8)) self.lineEditSynapsePort.setText(QtGui.QApplication.translate("Form", "N/A", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonSynapseServer.setText(QtGui.QApplication.translate("Form", "Connect", None, QtGui.QApplication.UnicodeUTF8)) self.labelChartTop.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">Raw EEG Waves</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.labelChartBottom.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">Brain Signals History</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.label.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Cantarell\'; font-size:11pt; font-weight:400; font-style:normal;\">\n" "<p align=\"center\" style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-family:\'Sans\'; font-size:10pt; font-weight:600;\">Current Data</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelControlConcentration.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p align=\"center\" style=\" margin-top:12px; margin-bottom:12px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-size:9pt;\">Concentration</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonControlConcentrationEnable.setToolTip(QtGui.QApplication.translate("Form", "enable/disable concentration processing", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonControlConcentrationEnable.setText(QtGui.QApplication.translate("Form", "Enabled", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelControlRelaxation.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p align=\"center\" style=\" margin-top:12px; margin-bottom:12px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-size:9pt;\">Relaxation</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonControlRelaxationEnable.setToolTip(QtGui.QApplication.translate("Form", "enable/disable relaxation processing", None, QtGui.QApplication.UnicodeUTF8)) self.pushButtonControlRelaxationEnable.setText(QtGui.QApplication.translate("Form", "Enabled", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelControlConnectionLevel.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Sans\'; font-size:10pt; font-weight:400; font-style:normal;\">\n" "<p align=\"center\" style=\" margin-top:12px; margin-bottom:12px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\">Connection Level</p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelSessionTimeTitle.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Cantarell\'; font-size:11pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">Connection Time</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelSessionTime.setText(QtGui.QApplication.translate("Form", "00:00:00", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelPacketsReceivedTitle.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Cantarell\'; font-size:11pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">Packets Received</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelPacketsReceived.setText(QtGui.QApplication.translate("Form", "0", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelPacketsDroppedTitle.setText(QtGui.QApplication.translate("Form", "<!DOCTYPE HTML PUBLIC \"-//W3C//DTD HTML 4.0//EN\" \"http://www.w3.org/TR/REC-html40/strict.dtd\">\n" "<html><head><meta name=\"qrichtext\" content=\"1\" /><style type=\"text/css\">\n" "p, li { white-space: pre-wrap; }\n" "</style></head><body style=\" font-family:\'Cantarell\'; font-size:11pt; font-weight:400; font-style:normal;\">\n" "<p style=\" margin-top:0px; margin-bottom:0px; margin-left:0px; margin-right:0px; -qt-block-indent:0; text-indent:0px;\"><span style=\" font-weight:600;\">Packets Dropped</span></p></body></html>", None, QtGui.QApplication.UnicodeUTF8)) self.textLabelPacketsDropped.setText(QtGui.QApplication.translate("Form", "0", None, QtGui.QApplication.UnicodeUTF8)) from PySide import QtWebKit
agpl-3.0
-5,169,383,353,182,562,000
77.328829
289
0.807637
false
pcaro90/Python-AES
ProgressBar.py
1
1773
#!/usr/bin/env python # -*- coding: utf-8 -*- # --------------------------------------------------- # Copyright (c) 2013 Pablo Caro. All Rights Reserved. # Pablo Caro <[email protected]> - http://pcaro.es/ # ProgressBar.py # --------------------------------------------------- import sys class ProgressBar: def __init__(self, min=0, max=100, width=60, charset='[=]'): self.min = min self.max = max self.width = width self.current = min self.percent = 0.0 self.int_percent = 0 if len(charset) != 3: charset = '[=]' self.charset = charset self.bar = '' self.used = -1 self.int_percent_change = False def update(self, current): self.current = current self.percent = (float(self.current-self.min)/(self.max-self.min))*100.0 int_percent = int(self.percent) if int_percent != self.int_percent: self.int_percent_change = True self.int_percent = int_percent self.__generate_bar__() if self.int_percent_change: self.int_percent_change = False return True else: return False def show(self): sys.stdout.write(str(self)) sys.stdout.flush() def __str__(self): return self.bar def __generate_bar__(self): self.used = int((float(self.current-self.min)/(self.max-self.min)) * (self.width-6)) center = self.charset[1] * self.used self.bar = (self.charset[0] + center + self.charset[2] + " " + str(self.int_percent) + '%' + '\r') def main(): pass if __name__ == '__main__': main()
bsd-3-clause
3,480,556,556,003,482,000
24.462687
79
0.478285
false
kefatong/ops
app/models.py
1
24843
# -*- coding:utf-8 -*- __author__ = 'eric' import hashlib from datetime import datetime from itsdangerous import TimedJSONWebSignatureSerializer as Serializer from werkzeug.security import generate_password_hash, check_password_hash from flask.ext.login import UserMixin from flask import current_app, request, url_for from . import db from . import login_manager class Permission: ADMINISTER = 0x001 DEVICE_LOOK = 0x002 DEVICE_EDIT = 0x004 DEVICE_DEL = 0x008 RACK_LOOK = 0x010 RACK_EDIT = 0x020 RACK_DEL = 0x040 IDC_LOOK = 0x080 IDC_EDIT = 0x100 IDC_DEL = 0x200 VIRTMACHINE_LOOK = 0x400 VIRTMACHINE_EDIT = 0x800 VIRTMACHINE_DEL = 0x1000 NETWORKDEVICE_LOOK = 0x2000 NETWORKDEVICE_EDIT = 0x4000 NETWORKDEVICE_DEL = 0x8000 DATABASE_LOOK = 0x10000 DATABASE_EDIT = 0x20000 DATABASE_DEL = 0x40000 SOFTWARE_LOOK = 0x80000 SOFTWARE_EDIT = 0x100000 SOFTWARE_DEL = 0x200000 DEVICEPERMISSION_LOOK = 0x400000 DEVICEPERMISSION_EDIT = 0x800000 DEVICEPERMISSION_DEL = 0x1000000 ASSET_LOOK = 0x2000000 ASSET_EDIT = 0x4000000 ASSET_DEL = 0x8000000 USER_LOOK = 0x10000000 USER_EDIT = 0x20000000 USER_DEL = 0x40000000 class Role(db.Model): __tablename__ = 'roles' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64), unique=True) default = db.Column(db.Boolean, index=True, default=False) permissions = db.Column(db.Integer) users = db.relationship('User', backref='role', lazy='dynamic') def to_json(self): json_role = { 'url': self.id, 'name': self.name, 'default': self.default, 'permissions': self.permissions, 'users': self.users, } return json_role @staticmethod def insert_roles(): roles = { 'User': (Permission.DEVICE_LOOK, True), 'manager': (Permission.USER_EDIT | Permission.DEVICE_LOOK | Permission.DEVICE_EDIT, False), 'Administrator': (Permission.USER_EDIT | Permission.DEVICE_LOOK | Permission.DEVICE_EDIT | Permission.DEVICE_DEL | Permission.ADMINISTER, False) } for r in roles: role = Role.query.filter_by(name=r).first() if role is None: role = Role(name=r) role.permissions = roles[r][0] role.default = roles[r][1] db.session.add(role) db.session.commit() def __repr__(self): return '<Role %r>' % self.name class User(UserMixin, db.Model): __tablename__ = 'users' id = db.Column(db.Integer, primary_key=True) email = db.Column(db.String(64), unique=True, index=True) # Email Address username = db.Column(db.String(64), unique=True, index=True) # Username password_hash = db.Column(db.String(128)) # password Md5 Hash role_id = db.Column(db.Integer, db.ForeignKey('roles.id')) # Role 鍏宠仈 Role table name = db.Column(db.String(64)) # 鐪熷疄濮撳悕 location = db.Column(db.String(64)) # 鍦板潃 position = db.Column(db.String(64)) # 职位 about_me = db.Column(db.Text()) # 鍏充簬鎴� phone = db.Column(db.String(11)) # 手机号码 qq = db.Column(db.String(13)) # QQ号码 member_since = db.Column(db.DateTime(), default=datetime.utcnow) # 娉ㄥ唽鏃堕棿 last_seen = db.Column(db.DateTime(), default=datetime.utcnow) # 鏈�鍚庣櫥褰曟椂闂� confirmed = db.Column(db.Boolean, default=False) # 璐︽埛鐘舵�� avatar_hash = db.Column(db.String(32)) # 澶村儚 logs = db.relationship('Logger', backref='user', lazy='dynamic') def to_json(self): json_user = { 'url': self.id, 'email': self.email, 'username': self.username, 'password_hash': self.password_hash, 'role': self.role, 'name': self.name, 'location': self.location, 'position': self.position, 'about_me': self.about_me, 'phone': self.phone, 'qq': self.qq, 'member_sine': self.member_since, 'last_seen': self.last_seen, } return json_user def __init__(self, **kwargs): super(User, self).__init__(**kwargs) if self.role is None: if self.email == current_app.config.get('FLASK_ADMIN', None): self.role = Role.query.filter_by(permissions=0xff).first() if self.role is None: self.role = Role.query.filter_by(default=True).first() if self.email is not None and self.avatar_hash is None: self.avatar_hash = hashlib.md5(self.email.encode('UTF-8')).hexdigest() @staticmethod def insert_admin_user(): r = Role() r.insert_roles() adminRole = Role.query.all()[-1] u = User.query.filter_by(username='administrator').first() if u is None: u = User() u.name = 'Admin' u.email = '[email protected]' u.username = 'administrator' u.password = '123456' u.confirmed = True u.role = adminRole db.session.add(u) db.session.commit() @staticmethod def generate_fake(count=1000): from sqlalchemy.exc import IntegrityError from random import seed import forgery_py seed() for i in range(count): u = User(email=forgery_py.internet.email_address(), username=forgery_py.internet.user_name(True), password=forgery_py.lorem_ipsum.word(), confirmed=True, name=forgery_py.name.full_name(), location=forgery_py.address.city(), position=forgery_py.lorem_ipsum.sentence(), about_me=forgery_py.lorem_ipsum.sentence(), member_since=forgery_py.date.date(True)) db.session.add(u) try: db.session.commit() except IntegrityError: print "db commit email : {0} Error".format(u.email) db.session.rollback() def gravatar(self, size=100, default='identicon', rating='g'): if request.is_secure: url = 'https://secure.gravatar.com/avatar' else: url = 'http://secure.gravatar.com/avatar' hash = self.avatar_hash or hashlib.md5(self.email.encode('UTF-8')).hexdigest() return '{url}/{hash}?s={size}&d={default}&r={rating}'.format(url=url, hash=hash, size=size, default=default, rating=rating) def ping(self): self.last_seen = datetime.utcnow() db.session.add(self) def can(self, permissions): return self.role is not None and (self.role.permissions & permissions) == permissions def is_administrator(self): return self.can(Permission.ADMINISTER) @property def password(self): raise AttributeError('password is not a readable attribute') @password.setter def password(self, password): self.password_hash = generate_password_hash(password) def verify_password(self, password): return check_password_hash(self.password_hash, password) def generate_confirmation_token(self, expiration=3600): s = Serializer(current_app.config['SECRET_KEY'], expires_in=expiration) return s.dumps({'confirm': self.id}) def confirm(self, token): s = Serializer(current_app.config['SECRET_KEY']) try: data = s.loads(token) except: return False if data.get('confirm', None) != self.id: return False self.confirmed = True print self.confirmed db.session.add(self) return True def generate_reset_token(self, expiration=3600): s = Serializer(current_app.config['SECRET_KEY'], expires_in=expiration) return s.dumps({'reset': self.id}) def reset_password(self, token, new_password): s = Serializer(current_app.config['SECRET_KEY']) try: data = s.loads(token) except: return False if data.get('reset', None) != self.id: return False self.password = new_password db.session.add(self) return True def generate_change_email_token(self, new_email, expiration=3600): s = Serializer(current_app.config['SECRET_KEY'], expires_in=expiration) return s.dumps({'change_email': self.id, 'new_email': new_email}) def change_email(self, token): s = Serializer(current_app.config['SECRET_KEY']) try: data = s.loads(token) except: return False if data.get('change_email') != self.id: return False new_email = data.get('new_email', None) if new_email is None: return False if self.query.filter_by(email=new_email).first() is not None: return False self.email = new_email self.avatar_hash = hashlib.md5(self.email.encode('UTF-8')).hexdigest() db.session.add(self) return True def generate_auth_token(self, expiration=3600): s = Serializer(current_app.config['SECRET_KEY'], expires_in=expiration) return s.dumps({'id': self.id}) @staticmethod def verify_auth_token(token): s = Serializer(current_app.config['SECRET_KEY']) try: data = s.loads(token) except: return None return User.query.get(data['id']) def __repr__(self): return '<User %r>' % self.username class DevicePower(db.Model): __tablename__ = 'devicePowers' id = db.Column(db.Integer, primary_key=True) type = db.Column(db.Integer) enabled = db.Column(db.Boolean, default=False) ip = db.Column(db.String(64)) # 杩滄帶鍗P鍦板潃 user = db.Column(db.String(64)) password_hash = db.Column(db.String(256)) powerid = db.Column(db.String(256)) device_id = db.Column(db.ForeignKey('devices.id')) isdelete = db.Column(db.Boolean, default=False) # 鏄惁鍒犻櫎 remarks = db.Column(db.Text) # 澶囨敞 instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 def generate_password_token(self, password): from itsdangerous import JSONWebSignatureSerializer as Serializer s = Serializer(current_app.config['SECRET_KEY']) return s.dumps({'confirm': password}) @property def password(self): raise AttributeError('password is not a readable attribute') @password.setter def password(self, password): self.password_hash = self.generate_password_token(password) def to_json(self): json_power = { 'url': self.id, 'type': self.type, 'enabled': self.enabled, 'ip': self.ip, 'user': self.user, 'password': self.password_hash, 'powerid': self.powerid, 'device_id': self.device_id, } return json_power def __repr__(self): return '<DevicePower %r>' % self.id DeviceRelationshipDeviceGroup = db.Table('DeviceRelationshipDeviceGroup', db.Column('deviceGroup_id', db.Integer, db.ForeignKey('deviceGroup.id')), db.Column('device_id', db.Integer, db.ForeignKey('devices.id')), ) class DeviceGroup(db.Model): __tablename__ = 'deviceGroup' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) business = db.Column(db.String(64)) # 所属业务 devices = db.relationship('Device', secondary=DeviceRelationshipDeviceGroup, backref=db.backref('DeviceGroup', lazy='dynamic'), lazy='dynamic') isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 remarks = db.Column(db.Text) # 澶囨敞 class Device(db.Model): __tablename__ = 'devices' id = db.Column(db.Integer, primary_key=True) device_id = db.Column(db.Integer, unique=True, index=True) hostname = db.Column(db.String(64)) # Hostname ip = db.Column(db.String(64)) an = db.Column(db.String(64), unique=True, index=True) sn = db.Column(db.String(64), unique=True, index=True) # SN 璁惧搴忓垪鍙� os = db.Column(db.String(64)) # os绫诲瀷 manufacturer = db.Column(db.String(64)) # 鐢熶骇鍟� brand = db.Column(db.String(64)) # 鍝佺墝 model = db.Column(db.String(64)) # 鍨嬪彿 cpumodel = db.Column(db.String(64)) # CPU 鍨嬪彿 cpucount = db.Column(db.Integer) # CPU 鏍告暟 memsize = db.Column(db.Integer) # 鍐呭瓨瀹归噺 disksize = db.Column(db.String(64)) business = db.Column(db.String(64)) # 所属业务 powerstatus = db.Column(db.Integer) # 电源状态 onstatus = db.Column(db.Integer) # 浣跨敤鐘舵�� usedept = db.Column(db.String(64)) # 浣跨敤閮ㄩ棬 usestaff = db.Column(db.String(64)) # 閮ㄩ棬浣跨敤浜� mainuses = db.Column(db.String(128)) # 涓昏鐢ㄩ�� isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 remarks = db.Column(db.Text) # 澶囨敞 def __repr__(self): return '<Device %r>' % self.hostname class ModuleClass(db.Model): __tablename__ = 'moduleClass' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) separator = db.Column(db.String(64), default='\t') isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 remarks = db.Column(db.Text) # 澶囨敞 def __repr__(self): return '<ModuleClass %r>' % self.name class TaskClass(db.Model): __tablename__ = 'taskClass' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) module_id = db.Column(db.Integer, db.ForeignKey('moduleClass.id')) separator = db.Column(db.String(64), default=' ') isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 remarks = db.Column(db.Text) # 澶囨敞 def __repr__(self): return '<TaskClass %r>' % self.name TaskRelationshipTaskGroup = db.Table('TaskRelationshipTaskGroup', db.Column('deviceTaskGroup_id', db.Integer, db.ForeignKey('deviceTaskGroup.id')), db.Column('deviceTask_id', db.Integer, db.ForeignKey('deviceTasks.id')), db.Column('PQ', db.Integer) ) class DeviceTaskGroup(db.Model): __tablename__ = 'deviceTaskGroup' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) enabled = db.Column(db.Boolean) # 是否启用 type = db.Column(db.Integer) tasks = db.relationship('DeviceTasks', secondary=TaskRelationshipTaskGroup, backref=db.backref('taskGroup', lazy='dynamic'), lazy='dynamic') isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 褰曞叆浜� inputtime = db.Column(db.DateTime, default=datetime.now) # 褰曞叆鏃堕棿 remarks = db.Column(db.Text) # 澶囨敞 def __repr__(self): return '<TaskScripts %r>' % self.name class DeviceTasks(db.Model): __tablename__ = 'deviceTasks' id = db.Column(db.Integer, primary_key=True) taskname = db.Column(db.String(64)) # 任务名称 scriptname = db.Column(db.String(256)) # 脚本名称 type = db.Column(db.Integer) # 脚本类型 python shell playbook perl arch = db.Column(db.Integer) # 系统架构 避免脚本运行出错 md5code = db.Column(db.String(128)) # 脚本md5码 防止被修改 path = db.Column(db.String(256)) # 脚本uuid extra_vars = db.Column(db.Text) version = db.Column(db.String(20)) # 脚本版本 enabled = db.Column(db.Boolean) # 启用 isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 def __repr__(self): return '<TaskScripts %r>' % self.taskname class System(db.Model): __tablename__ = 'System' id = db.Column(db.Integer, primary_key=True) device_id = db.Column(db.Integer) # 设备id 通过cmdb读取设备 an = db.Column(db.String(64)) sn = db.Column(db.String(64)) ip = db.Column(db.String(20)) # 设备IP地址 选择设备后自动获取设备IP 在正式安装时判断是否与设备一致 hostname = db.Column(db.String(64)) power_ip = db.Column(db.String(32)) os_version = db.Column(db.String(64)) # 系统版本 type = db.Column(db.Integer) # 部署方式 IPMI DHCP post = db.Column(db.Integer, db.ForeignKey('deviceTaskGroup.id')) # 任务列表 安装系统后需要执行的 status = db.Column(db.Integer, default=1) isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 def __repr__(self): return '<TaskScripts %r>' % self.sn class ComplianceTasks(db.Model): __tablename__ = 'Compliance' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) deviceGroup = db.Column(db.Integer, db.ForeignKey('deviceGroup.id')) taskGroup = db.Column(db.Integer, db.ForeignKey('deviceTaskGroup.id')) status = db.Column(db.Integer, default=1) enabled = db.Column(db.Boolean) isdelete = db.Column(db.Boolean) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 class ComplianceRecord(db.Model): __tablename__ = 'ComplianceRecord' id = db.Column(db.Integer, primary_key=True) SoftwareRelationshipDevice = db.Table('SoftwareRelationshipDevice', db.Column('SoftwareDistribution_id', db.Integer, db.ForeignKey('SoftwareDistribution.id')), db.Column('device_id', db.Integer, db.ForeignKey('devices.id')) ) class SoftwareDistribution(db.Model): __tablename__ = 'SoftwareDistribution' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64), unique=True) devices = db.relationship('Device', secondary=SoftwareRelationshipDevice, lazy='dynamic') taskGroup = db.Column(db.Integer, db.ForeignKey('deviceTaskGroup.id')) type = db.Column(db.Integer) status = db.Column(db.Integer) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 ContrastTaskRelationshipContrastFilesOrDirectory = db.Table('ContrastTaskRelationshipContrastFilesOrDirectory', db.Column('ContrastTask_id', db.Integer, db.ForeignKey('ContrastTasks.id')), db.Column('ContrastFilesOrDirectory', db.Integer, db.ForeignKey('ContrastFilesOrDirectory.id')) ) class ContrastTasks(db.Model): __tablename__ = 'ContrastTasks' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) fileOrDirectory = db.relationship('ContrastFilesOrDirectory', secondary=ContrastTaskRelationshipContrastFilesOrDirectory, backref=db.backref('tasks', lazy='dynamic'), lazy='dynamic') type = db.Column(db.Integer) enabled = db.Column(db.Boolean) status = db.Column(db.Integer) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 class ContrastFilesOrDirectory(db.Model): __tablename__ = 'ContrastFilesOrDirectory' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) device_id = db.Column(db.ForeignKey('devices.id')) type = db.Column(db.Integer) # 对比类型, 文件或目录 path = db.Column(db.String(512)) enabled = db.Column(db.String(64)) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 class ContrastResults(db.Model): __tablename__ = 'ContrastResults' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64)) task = db.Column(db.ForeignKey('ContrastTasks.id')) template = db.Column(db.String(512)) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 ApplicationOfReleaseRelationshipDevice = db.Table('ApplicationOfReleaseRelationshipDevice', db.Column('ApplicationOfRelease_id', db.Integer, db.ForeignKey('ApplicationOfRelease.id')), db.Column('device_id', db.Integer, db.ForeignKey('devices.id')) ) class ApplicationOfRelease(db.Model): __tablename__ = 'ApplicationOfRelease' id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(64), unique=True) devices = db.relationship('Device', secondary=ApplicationOfReleaseRelationshipDevice, lazy='dynamic') taskGroup = db.Column(db.Integer, db.ForeignKey('deviceTaskGroup.id')) type = db.Column(db.Integer) status = db.Column(db.Integer) instaff = db.Column(db.String(64)) # 录入人 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 class histroyCommands(db.Model): __tablename__ = 'histroyCommands' id = db.Column(db.Integer, primary_key=True) user_id = db.Column(db.Integer, db.ForeignKey('users.id')) command = db.Column(db.String(512)) # 命令记录 inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 remarks = db.Column(db.Text) # 备注 class pushCommandLogger(db.Model): __tablename__ = 'pushCommandLogger' id = db.Column(db.Integer, primary_key=True) device_id = db.Column(db.Integer, db.ForeignKey('devices.id')) command = db.Column(db.Text) inputtime = db.Column(db.DateTime, default=datetime.now) # 录入时间 def __repr__(self): return '<command %r>' % self.command class Logger(db.Model): __tablename__ = 'logs' id = db.Column(db.Integer, primary_key=True) user_id = db.Column(db.Integer, db.ForeignKey('users.id')) logtime = db.Column(db.DateTime, default=datetime.now()) content = db.Column(db.String(256)) # action [ 1: add , 2: edit, 3: del ] action = db.Column(db.String(32)) logobjtype = db.Column(db.String(64)) logobj_id = db.Column(db.Integer) def __repr__(self): return '<Logs %r>' % self.user_id @login_manager.user_loader def load_user(user_id): return User.query.get(int(user_id))
apache-2.0
-4,587,166,439,716,651,500
34.864823
133
0.588358
false
gbjuno/coreprogramming
chapter7/7-6.py
1
1400
#!/usr/bin/env python def stockDataSort(stockData,userInput): stockDataAfterSort = {} if userInput == None: print "input wrong items" exit(1) else: for i in stockData: stockDataAfterSort[i[userInput]] = i print "-----------------------------------" for dataItem in sorted(stockDataAfterSort): print "%s : %s" % (dataItem,stockDataAfterSort[dataItem]) print "-----------------------------------" return stockDataAfterSort def main(): stockData = [ ['1000',1.6,1.7,1000], ['3023',101.5,0,0], ['0032',300.1,298,300], ['2032',30.2,40,60000] ] userInputMap = {"number":0,"currentPrice":1,"buyPrice":2,"stockNumber":3} userInput = "number" print "stockDataAfterSort is %s " % stockDataSort(stockData,userInputMap.get(userInput)) userInput = "currentPrice" print "stockDataAfterSort is %s " % stockDataSort(stockData,userInputMap.get(userInput)) userInput = "buyPrice" print "stockDataAfterSort is %s " % stockDataSort(stockData,userInputMap.get(userInput)) userInput = "stockNumber" print "stockDataAfterSort is %s " % stockDataSort(stockData,userInputMap.get(userInput)) userInput = "haha" print "stockDataAfterSort is %s " % stockDataSort(stockData,userInputMap.get(userInput)) if __name__ == '__main__': main()
mit
1,574,336,147,322,682,400
32.333333
92
0.604286
false
futurice/vdsm
vdsm/network/configurators/__init__.py
1
5484
# Copyright 2013-2014 Red Hat, Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA # # Refer to the README and COPYING files for full details of the license # import ConfigParser import logging from vdsm import netinfo from vdsm.config import config from vdsm.netconfpersistence import RunningConfig from . import libvirt from ..models import Bond, Bridge from ..sourceroute import StaticSourceRoute class RollbackIncomplete(Exception): pass class Configurator(object): def __init__(self, configApplier, inRollback=False): self.configApplier = configApplier self._inRollback = inRollback self._libvirtAdded = set() def __enter__(self): self.begin() return self def __exit__(self, type, value, traceback): if type is None: self.commit() elif self._inRollback: # If we failed the rollback transaction, the networking system # is in no good state and we fail hard logging.error('Failed rollback transaction last known good ' 'network. ERR=%s', exc_info=(type, value, traceback)) else: leftover = self.rollback() if leftover: raise RollbackIncomplete(leftover, type, value) def rollback(self): """ returns None when all the nets were successfully rolled back, a vdsm.netoconfpersistence.Config object with the not yet rolled back networks and bonds. """ # self.runningConfig will have all the changes that were applied before # we needed to rollback. return RunningConfig().diffFrom(self.runningConfig) def flush(self): libvirt.flush() def configureBridge(self, bridge, **opts): raise NotImplementedError def configureVlan(self, vlan, **opts): raise NotImplementedError def configureBond(self, bond, **opts): raise NotImplementedError def editBonding(self, bond, _netinfo): raise NotImplementedError def configureNic(self, nic, **opts): raise NotImplementedError def removeBridge(self, bridge): raise NotImplementedError def removeVlan(self, vlan): raise NotImplementedError def removeBond(self, bonding): raise NotImplementedError def removeNic(self, nic): raise NotImplementedError def configureSourceRoute(self, routes, rules, device): raise NotImplementedError def removeSourceRoute(self, routes, rules, device): raise NotImplementedError def configureLibvirtNetwork(self, network, iface, qosInbound=None, qosOutbound=None): self.configApplier.createLibvirtNetwork(network, isinstance(iface, Bridge), iface.name, qosInbound=qosInbound, qosOutbound=qosOutbound) self._libvirtAdded.add(network) def removeLibvirtNetwork(self, network): self.configApplier.removeLibvirtNetwork(network) def _addSourceRoute(self, netEnt): ip = netEnt.ipConfig # bootproto is None for both static and no bootproto if ip.bootproto != 'dhcp' and netEnt.master is None: logging.debug("Adding source route %s, %s, %s, %s" % (netEnt.name, ip.ipaddr, ip.netmask, ip.gateway)) StaticSourceRoute(netEnt.name, self).\ configure(ip.ipaddr, ip.netmask, ip.gateway) def _removeSourceRoute(self, netEnt, sourceRouteClass): if netEnt.ipConfig.bootproto != 'dhcp' and netEnt.master is None: logging.debug("Removing source route for device %s", netEnt.name) sourceRouteClass(netEnt.name, self).remove() def _setNewMtu(self, iface, ifaceVlans): """ Update an interface's MTU when one of its users is removed. :param iface: interface object (bond or nic device) :type iface: NetDevice instance :param ifaceVlans: vlan devices using the interface 'iface' :type ifaceVlans: iterable :return mtu value that was applied """ ifaceMtu = netinfo.getMtu(iface.name) maxMtu = netinfo.getMaxMtu(ifaceVlans, None) if maxMtu and maxMtu < ifaceMtu: if isinstance(iface, Bond): self.configApplier.setBondingMtu(iface.name, maxMtu) else: self.configApplier.setIfaceMtu(iface.name, maxMtu) return maxMtu def getEthtoolOpts(name): try: opts = config.get('vars', 'ethtool_opts.' + name) except ConfigParser.NoOptionError: opts = config.get('vars', 'ethtool_opts') return opts
gpl-2.0
1,205,282,231,433,251,600
33.708861
79
0.642232
false
dhermes/ncaa-bracket-scenarios
run_all_scenarios.py
1
1579
from __future__ import print_function import itertools import json import pickle from game_tree_classes import WinnerOf import utils with open(utils.SWEET16_PICKLE, 'r') as fh: SLOTS_BEFORE = pickle.load(fh) def complete_bracket(game_slots, choice_slots, choices): result = game_slots.copy() for slot_id, choice_val in zip(choice_slots, choices): winner_of = result.get_slot(slot_id) if choice_val not in (winner_of.game_slot1, winner_of.game_slot2): raise ValueError('Choice does not match available.') winning_team = result.get_slot(choice_val) result.reset_slot(slot_id, winning_team) if not result.complete: raise ValueError('Expected bracket to be complete.') return result.reduced def main(): choice_slots = [] choice_vals = [] for slot_id in xrange(127): value = SLOTS_BEFORE.get_slot(slot_id) if isinstance(value, WinnerOf): choice_slots.append(slot_id) choice_vals.append((value.game_slot1, value.game_slot2)) msg = '{:d} choices left'.format(len(choice_slots)) print(msg) reduced_vals = [] for choice_tuple in itertools.product(*choice_vals): reduced_vals.append( complete_bracket(SLOTS_BEFORE, choice_slots, choice_tuple)) filename = utils.REDUCED_SCENARIOS with open(filename, 'w') as fh: json.dump(reduced_vals, fh, indent=2, sort_keys=True, separators=(',', ': ')) msg = 'Created {}'.format(filename) print(msg) if __name__ == '__main__': main()
apache-2.0
-6,612,150,762,153,403,000
28.240741
74
0.639012
false
Gaia3D/QGIS
python/plugins/processing/algs/qgis/DeleteHoles.py
1
2935
# -*- coding: utf-8 -*- """ *************************************************************************** DeleteHoles.py --------------------- Date : April 2015 Copyright : (C) 2015 by Etienne Trimaille *************************************************************************** * * * This program is free software; you can redistribute it and/or modify * * it under the terms of the GNU General Public License as published by * * the Free Software Foundation; either version 2 of the License, or * * (at your option) any later version. * * * *************************************************************************** """ __author__ = 'Etienne Trimaille' __date__ = 'April 2015' __copyright__ = '(C) 2015, Etienne Trimaille' # This will get replaced with a git SHA1 when you do a git archive __revision__ = '$Format:%H$' from qgis.core import QgsFeature, QgsGeometry from processing.core.GeoAlgorithm import GeoAlgorithm from processing.core.parameters import ParameterVector from processing.core.outputs import OutputVector from processing.tools import dataobjects, vector class DeleteHoles(GeoAlgorithm): INPUT = 'INPUT' OUTPUT = 'OUTPUT' def defineCharacteristics(self): self.name = 'Delete holes' self.group = 'Vector geometry tools' self.addParameter(ParameterVector(self.INPUT, self.tr('Input layer'), [ParameterVector.VECTOR_TYPE_POLYGON])) self.addOutput(OutputVector(self.OUTPUT, self.tr('Output'))) def processAlgorithm(self, progress): layer = dataobjects.getObjectFromUri( self.getParameterValue(self.INPUT)) writer = self.getOutputFromName(self.OUTPUT).getVectorWriter( layer.pendingFields(), layer.wkbType(), layer.crs()) features = vector.features(layer) count = len(features) total = 100.0 / float(count) feat = QgsFeature() for count, f in enumerate(features): geometry = f.geometry() if geometry.isMultipart(): multi_polygon = geometry.asMultiPolygon() for polygon in multi_polygon: for ring in polygon[1:]: polygon.remove(ring) geometry = QgsGeometry.fromMultiPolygon(multi_polygon) else: polygon = geometry.asPolygon() for ring in polygon[1:]: polygon.remove(ring) geometry = QgsGeometry.fromPolygon(polygon) feat.setGeometry(geometry) feat.setAttributes(f.attributes()) writer.addFeature(feat) progress.setPercentage(int(count * total)) del writer
gpl-2.0
-8,316,699,685,901,509,000
35.246914
75
0.527768
false
smira/spamfighter
spamfighter/core/null_partner.py
1
3199
# -*- coding: utf-8 -*- # # SpamFighter, Copyright 2008, 2009 NetStream LLC (http://netstream.ru/, [email protected]) # # This file is part of SpamFighter. # # SpamFighter is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # SpamFighter is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with SpamFighter. If not, see <http://www.gnu.org/licenses/>. # """ Модуль авторизации партнеров без логинов/паролей (на доверии). """ from zope.interface import implements from twisted.internet import defer from spamfighter.interfaces import IPartner, IPartnerAuthorizer from spamfighter.core.partner import PartnerAuthorizationFailedError from spamfighter.core.domain import getDefaultDomain, BaseDomain from spamfighter.plugin import loadPlugin, IDefaultDomainProvider from spamfighter.utils import config class NullPartner(object): """ Партнер, авторизованный без логина/пароля (на доверии). @ivar domain: корневой домен партнера @type domain: L{BaseDomain} """ implements(IPartner) def __init__(self): """ Конструктор. """ domainProvider = loadPlugin(IDefaultDomainProvider, config.plugins.domain.null_partner_domain_provider) self.domain = domainProvider.getDefaultDomain() def rootDomain(self): """ Получить корневой домен партнера. @return: Deferred, корневой домен (L{IDomain}) @rtype: C{twisted.internet.defer.Deferred} """ return defer.succeed(self.domain) class NullPartnerAuthorizer(object): """ Провайдер авторизации партнеров без логина/пароля (на доверии). В этой ситуации доступ к СпамоБорцу ограничен с помощью других средств (HTTP-proxy, firewall). @ivar partner: единственный партнер, который обеспечивает весь доступ @type partner: L{NullPartner} """ implements(IPartnerAuthorizer) def __init__(self): """ Конструктор. """ self.partner = NullPartner() def authorize(self, partner_info): """ Выполнить авторизацию партнера. @param partner_info: информация о партнере @return: Deferred, партнер (L{IPartner}) @rtype: C{twisted.internet.defer.Deferred} """ if partner_info is not None: return defer.fail(PartnerAuthorizationFailedError()) return defer.succeed(self.partner)
gpl-3.0
-2,622,230,818,337,814,000
30.840909
111
0.70414
false
lowitty/eeep
insertdb/insertmodeldata.py
1
6340
#encoding=utf-8 import MySQLdb, os from django.core.checks.registry import Tags def insertDomain(db): cursor = db.cursor() sql = "insert into %(table)s (%(para)s) values ('%(value)s')" insertValues = {'table' : 'quotation_domain', 'para' : 'name', 'value' : 'CD'} f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + 'domain', 'r+') for line in f.readlines(): line = line.strip() if('' != line): insertValues['value'] = line exesql = sql % insertValues cursor.execute(exesql) db.commit() db.close() def insertSubDomain(db): cursor = db.cursor() sql = "insert into %(table)s (%(para)s) values ('%(value)s')" insertValues = {'table' : 'quotation_domain', 'para' : 'name', 'value' : 'CD'} insertValues['table'] = 'quotation_subdomain' f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + 'subdomain', 'r+') for line in f.readlines(): line = line.strip() if('' != line): insertValues['value'] = line exesql = sql % insertValues cursor.execute(exesql) db.commit() db.close() def insertRegion(db, tableName, valueTag, fileName): cursor = db.cursor() sql = "insert into %(table)s (%(valueTag)s) values ('%(value)s')" insertValues = {'table' : tableName, 'valueTag' : valueTag, 'value' : 'xxxxx'} #print sql % insertValues f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + fileName, 'r+') for line in f.readlines(): line = line.strip() if('' != line): para = line.split('**') if(len(para) > 1): insertValues['value'] = para[0].strip() cursor.execute(sql % insertValues) db.commit() db.close() def insertValuesWithForignKey(db, table, tagValue, tagForKey, f_table, f_tagvalue, fileName = 'unitid'): cursor = db.cursor() sql = "insert into %(table)s (" + tagValue + "," + tagForKey + ") values ('%(" + tagValue + ")s', %(" + tagForKey + ")s)" insertValues = {'table' : table, tagValue : 'OMS CD', tagForKey : 1} f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + fileName, 'r+') f_id = -1 exeTimes = 0 for line in f.readlines(): exeTimes += 1 line = line.strip() if('' != line): para = line.split('**') if(len(para) > 1): f_name = para[0].strip() cursor.execute("select id from %s where %s='%s'" % (f_table, f_tagvalue, f_name)) f_id = cursor.fetchone()[0] insertValues[tagValue] = para[1].strip().replace('\'', "\\'") insertValues[tagForKey] = f_id print sql % insertValues else: insertValues[tagValue] = para[0].strip().replace('\'', "\\'") insertValues[tagForKey] = f_id print sql % insertValues cursor.execute(sql % insertValues) if(exeTimes % 10 == 0): db.commit() #pass db.commit() db.close() def insertWorkcenter(db, tableName, fileName, *tags): if(4 != len(tags)): return False else: cursor = db.cursor() sql = "insert into %(tableName)s (" + tags[0] + "," + tags[1] + "," + tags[2] + "," + tags[3] + ") values ('%(" + tags[0] + ")s','%(" + tags[1] + ")s','%(" + tags[2] + ")s','%("+ tags[3] +")s')".encode('utf-8') insertDatas = { 'tableName' : tableName, tags[0] : '', tags[1] : '', tags[2] : '', tags[3] : '' } f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + fileName, 'r+') cIndex = 0 for line in f.readlines(): cIndex += 1 if('' != line): para = line.split('**') if(len(para) > 3): insertDatas[tags[0]] = para[0].strip().replace("\'", "\\'").encode('utf-8') insertDatas[tags[1]] = para[1].strip().replace("\'", "\\'").encode('utf-8') insertDatas[tags[2]] = para[2].strip().replace("\'", "\\'").encode('utf-8') insertDatas[tags[3]] = para[3].strip().replace("\'", "\\'").encode('utf-8') #print (sql % insertDatas).encode('utf-8') cursor.execute((sql % insertDatas).encode('utf-8')) if(cIndex % 10 == 0): db.commit() db.commit() db.close() def insertPostatus(db, fileName): cursor = db.cursor() sql = "insert into quotation_postatus (name) values ('%s')" f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + fileName, 'r+') for line in f.readlines(): line = line.strip() if('' != line): exesql = sql % line cursor.execute(exesql) db.commit() db.close() def insertOrderingcompany(db, fileName): cursor = db.cursor() sql = "insert into quotation_orderingcompany (name) values ('%s')" f = open(os.path.dirname(os.path.abspath(__file__)) + os.path.sep + fileName, 'r+') cIndex = 0 for line in f.readlines(): cIndex += 1 line = line.strip() if('' != line): exesql = sql % line #print exesql cursor.execute(exesql) if( 0 == cIndex % 10): db.commit() db.commit() db.close() if __name__ == '__main__': host = "localhost" passwd = "tatool" user = "tatool" dbname = "eeep" db = MySQLdb.connect(host=host, user=user, passwd=passwd, db=dbname) #insertDomain(db) #insertSubDomain(db) #insertValuesWithForignKey(db, 'quotation_unitid', 'name', 'domain_id', "quotation_domain", "name") #insertRegion(db, 'quotation_region', 'name', 'regionandcountry') #insertValuesWithForignKey(db, 'quotation_country', 'name', 'region_id', "quotation_region", "name", 'regionandcountry') #insertWorkcenter(db, 'quotation_workcenter', 'workcenter', 'number', 'descworkcenter', 'icrrbactivitytype', 'intracompanyactivitytyoe') #insertPostatus(db, 'postatus') insertOrderingcompany(db, 'orderingcompany')
mit
-8,395,665,122,548,171,000
37.90184
218
0.519085
false
xen0l/ansible
lib/ansible/modules/cloud/vmware/vmware_vm_facts.py
1
5877
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright: (c) 2015, Joseph Callen <jcallen () csc.com> # Copyright: (c) 2018, Ansible Project # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type ANSIBLE_METADATA = { 'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'community' } DOCUMENTATION = r''' --- module: vmware_vm_facts short_description: Return basic facts pertaining to a vSphere virtual machine guest description: - Return basic facts pertaining to a vSphere virtual machine guest. version_added: '2.0' author: - Joseph Callen (@jcpowermac) - Abhijeet Kasurde (@Akasurde) notes: - Tested on vSphere 5.5 and vSphere 6.5 requirements: - python >= 2.6 - PyVmomi options: vm_type: description: - If set to C(vm), then facts are gathered for virtual machines only. - If set to C(template), then facts are gathered for virtual machine templates only. - If set to C(all), then facts are gathered for all virtual machines and virtual machine templates. required: False default: 'all' choices: [ all, vm, template ] version_added: 2.5 extends_documentation_fragment: vmware.documentation ''' EXAMPLES = r''' - name: Gather all registered virtual machines vmware_vm_facts: hostname: '{{ vcenter_hostname }}' username: '{{ vcenter_username }}' password: '{{ vcenter_password }}' delegate_to: localhost register: vmfacts - debug: var: vmfacts.virtual_machines - name: Gather only registered virtual machine templates vmware_vm_facts: hostname: '{{ vcenter_hostname }}' username: '{{ vcenter_username }}' password: '{{ vcenter_password }}' vm_type: template delegate_to: localhost register: template_facts - debug: var: template_facts.virtual_machines - name: Gather only registered virtual machines vmware_vm_facts: hostname: '{{ vcenter_hostname }}' username: '{{ vcenter_username }}' password: '{{ vcenter_password }}' vm_type: vm delegate_to: localhost register: vm_facts - debug: var: vm_facts.virtual_machines ''' RETURN = r''' virtual_machines: description: dictionary of virtual machines and their facts returned: success type: dict ''' try: from pyVmomi import vim except ImportError: pass from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.vmware import PyVmomi, get_all_objs, vmware_argument_spec, _get_vm_prop class VmwareVmFacts(PyVmomi): def __init__(self, module): super(VmwareVmFacts, self).__init__(module) # https://github.com/vmware/pyvmomi-community-samples/blob/master/samples/getallvms.py def get_all_virtual_machines(self): """ Function to get all virtual machines and related configurations information """ virtual_machines = get_all_objs(self.content, [vim.VirtualMachine]) _virtual_machines = {} for vm in virtual_machines: _ip_address = "" summary = vm.summary if summary.guest is not None: _ip_address = summary.guest.ipAddress if _ip_address is None: _ip_address = "" _mac_address = [] all_devices = _get_vm_prop(vm, ('config', 'hardware', 'device')) if all_devices: for dev in all_devices: if isinstance(dev, vim.vm.device.VirtualEthernetCard): _mac_address.append(dev.macAddress) net_dict = {} vmnet = _get_vm_prop(vm, ('guest', 'net')) if vmnet: for device in vmnet: net_dict[device.macAddress] = dict() net_dict[device.macAddress]['ipv4'] = [] net_dict[device.macAddress]['ipv6'] = [] for ip_addr in device.ipAddress: if "::" in ip_addr: net_dict[device.macAddress]['ipv6'].append(ip_addr) else: net_dict[device.macAddress]['ipv4'].append(ip_addr) esxi_hostname = None if summary.runtime.host: esxi_hostname = summary.runtime.host.summary.config.name virtual_machine = { summary.config.name: { "guest_fullname": summary.config.guestFullName, "power_state": summary.runtime.powerState, "ip_address": _ip_address, # Kept for backward compatibility "mac_address": _mac_address, # Kept for backward compatibility "uuid": summary.config.uuid, "vm_network": net_dict, "esxi_hostname": esxi_hostname, } } vm_type = self.module.params.get('vm_type') is_template = _get_vm_prop(vm, ('config', 'template')) if vm_type == 'vm' and not is_template: _virtual_machines.update(virtual_machine) elif vm_type == 'template' and is_template: _virtual_machines.update(virtual_machine) elif vm_type == 'all': _virtual_machines.update(virtual_machine) return _virtual_machines def main(): argument_spec = vmware_argument_spec() argument_spec.update( vm_type=dict(type='str', choices=['vm', 'all', 'template'], default='all'), ) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=False) vmware_vm_facts = VmwareVmFacts(module) _virtual_machines = vmware_vm_facts.get_all_virtual_machines() module.exit_json(changed=False, virtual_machines=_virtual_machines) if __name__ == '__main__': main()
gpl-3.0
-3,481,050,722,794,838,500
31.832402
105
0.602008
false
googleapis/googleapis-gen
google/cloud/dialogflow/cx/v3beta1/dialogflow-cx-v3beta1-py/google/cloud/dialogflowcx_v3beta1/services/entity_types/client.py
1
38725
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from collections import OrderedDict from distutils import util import os import re from typing import Callable, Dict, Optional, Sequence, Tuple, Type, Union import pkg_resources from google.api_core import client_options as client_options_lib # type: ignore from google.api_core import exceptions as core_exceptions # type: ignore from google.api_core import gapic_v1 # type: ignore from google.api_core import retry as retries # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport import mtls # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore from google.auth.exceptions import MutualTLSChannelError # type: ignore from google.oauth2 import service_account # type: ignore from google.cloud.dialogflowcx_v3beta1.services.entity_types import pagers from google.cloud.dialogflowcx_v3beta1.types import entity_type from google.cloud.dialogflowcx_v3beta1.types import entity_type as gcdc_entity_type from google.protobuf import field_mask_pb2 # type: ignore from .transports.base import EntityTypesTransport, DEFAULT_CLIENT_INFO from .transports.grpc import EntityTypesGrpcTransport from .transports.grpc_asyncio import EntityTypesGrpcAsyncIOTransport class EntityTypesClientMeta(type): """Metaclass for the EntityTypes client. This provides class-level methods for building and retrieving support objects (e.g. transport) without polluting the client instance objects. """ _transport_registry = OrderedDict() # type: Dict[str, Type[EntityTypesTransport]] _transport_registry["grpc"] = EntityTypesGrpcTransport _transport_registry["grpc_asyncio"] = EntityTypesGrpcAsyncIOTransport def get_transport_class(cls, label: str = None, ) -> Type[EntityTypesTransport]: """Returns an appropriate transport class. Args: label: The name of the desired transport. If none is provided, then the first transport in the registry is used. Returns: The transport class to use. """ # If a specific transport is requested, return that one. if label: return cls._transport_registry[label] # No transport is requested; return the default (that is, the first one # in the dictionary). return next(iter(cls._transport_registry.values())) class EntityTypesClient(metaclass=EntityTypesClientMeta): """Service for managing [EntityTypes][google.cloud.dialogflow.cx.v3beta1.EntityType]. """ @staticmethod def _get_default_mtls_endpoint(api_endpoint): """Converts api endpoint to mTLS endpoint. Convert "*.sandbox.googleapis.com" and "*.googleapis.com" to "*.mtls.sandbox.googleapis.com" and "*.mtls.googleapis.com" respectively. Args: api_endpoint (Optional[str]): the api endpoint to convert. Returns: str: converted mTLS api endpoint. """ if not api_endpoint: return api_endpoint mtls_endpoint_re = re.compile( r"(?P<name>[^.]+)(?P<mtls>\.mtls)?(?P<sandbox>\.sandbox)?(?P<googledomain>\.googleapis\.com)?" ) m = mtls_endpoint_re.match(api_endpoint) name, mtls, sandbox, googledomain = m.groups() if mtls or not googledomain: return api_endpoint if sandbox: return api_endpoint.replace( "sandbox.googleapis.com", "mtls.sandbox.googleapis.com" ) return api_endpoint.replace(".googleapis.com", ".mtls.googleapis.com") DEFAULT_ENDPOINT = "dialogflow.googleapis.com" DEFAULT_MTLS_ENDPOINT = _get_default_mtls_endpoint.__func__( # type: ignore DEFAULT_ENDPOINT ) @classmethod def from_service_account_info(cls, info: dict, *args, **kwargs): """Creates an instance of this client using the provided credentials info. Args: info (dict): The service account private key info. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: EntityTypesClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_info(info) kwargs["credentials"] = credentials return cls(*args, **kwargs) @classmethod def from_service_account_file(cls, filename: str, *args, **kwargs): """Creates an instance of this client using the provided credentials file. Args: filename (str): The path to the service account private key json file. args: Additional arguments to pass to the constructor. kwargs: Additional arguments to pass to the constructor. Returns: EntityTypesClient: The constructed client. """ credentials = service_account.Credentials.from_service_account_file( filename) kwargs["credentials"] = credentials return cls(*args, **kwargs) from_service_account_json = from_service_account_file @property def transport(self) -> EntityTypesTransport: """Returns the transport used by the client instance. Returns: EntityTypesTransport: The transport used by the client instance. """ return self._transport @staticmethod def entity_type_path(project: str,location: str,agent: str,entity_type: str,) -> str: """Returns a fully-qualified entity_type string.""" return "projects/{project}/locations/{location}/agents/{agent}/entityTypes/{entity_type}".format(project=project, location=location, agent=agent, entity_type=entity_type, ) @staticmethod def parse_entity_type_path(path: str) -> Dict[str,str]: """Parses a entity_type path into its component segments.""" m = re.match(r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)/agents/(?P<agent>.+?)/entityTypes/(?P<entity_type>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_billing_account_path(billing_account: str, ) -> str: """Returns a fully-qualified billing_account string.""" return "billingAccounts/{billing_account}".format(billing_account=billing_account, ) @staticmethod def parse_common_billing_account_path(path: str) -> Dict[str,str]: """Parse a billing_account path into its component segments.""" m = re.match(r"^billingAccounts/(?P<billing_account>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_folder_path(folder: str, ) -> str: """Returns a fully-qualified folder string.""" return "folders/{folder}".format(folder=folder, ) @staticmethod def parse_common_folder_path(path: str) -> Dict[str,str]: """Parse a folder path into its component segments.""" m = re.match(r"^folders/(?P<folder>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_organization_path(organization: str, ) -> str: """Returns a fully-qualified organization string.""" return "organizations/{organization}".format(organization=organization, ) @staticmethod def parse_common_organization_path(path: str) -> Dict[str,str]: """Parse a organization path into its component segments.""" m = re.match(r"^organizations/(?P<organization>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_project_path(project: str, ) -> str: """Returns a fully-qualified project string.""" return "projects/{project}".format(project=project, ) @staticmethod def parse_common_project_path(path: str) -> Dict[str,str]: """Parse a project path into its component segments.""" m = re.match(r"^projects/(?P<project>.+?)$", path) return m.groupdict() if m else {} @staticmethod def common_location_path(project: str, location: str, ) -> str: """Returns a fully-qualified location string.""" return "projects/{project}/locations/{location}".format(project=project, location=location, ) @staticmethod def parse_common_location_path(path: str) -> Dict[str,str]: """Parse a location path into its component segments.""" m = re.match(r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)$", path) return m.groupdict() if m else {} def __init__(self, *, credentials: Optional[ga_credentials.Credentials] = None, transport: Union[str, EntityTypesTransport, None] = None, client_options: Optional[client_options_lib.ClientOptions] = None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, ) -> None: """Instantiates the entity types client. Args: credentials (Optional[google.auth.credentials.Credentials]): The authorization credentials to attach to requests. These credentials identify the application to the service; if none are specified, the client will attempt to ascertain the credentials from the environment. transport (Union[str, EntityTypesTransport]): The transport to use. If set to None, a transport is chosen automatically. client_options (google.api_core.client_options.ClientOptions): Custom options for the client. It won't take effect if a ``transport`` instance is provided. (1) The ``api_endpoint`` property can be used to override the default endpoint provided by the client. GOOGLE_API_USE_MTLS_ENDPOINT environment variable can also be used to override the endpoint: "always" (always use the default mTLS endpoint), "never" (always use the default regular endpoint) and "auto" (auto switch to the default mTLS endpoint if client certificate is present, this is the default value). However, the ``api_endpoint`` property takes precedence if provided. (2) If GOOGLE_API_USE_CLIENT_CERTIFICATE environment variable is "true", then the ``client_cert_source`` property can be used to provide client certificate for mutual TLS transport. If not provided, the default SSL client certificate will be used if present. If GOOGLE_API_USE_CLIENT_CERTIFICATE is "false" or not set, no client certificate will be used. client_info (google.api_core.gapic_v1.client_info.ClientInfo): The client info used to send a user-agent string along with API requests. If ``None``, then default info will be used. Generally, you only need to set this if you're developing your own client library. Raises: google.auth.exceptions.MutualTLSChannelError: If mutual TLS transport creation failed for any reason. """ if isinstance(client_options, dict): client_options = client_options_lib.from_dict(client_options) if client_options is None: client_options = client_options_lib.ClientOptions() # Create SSL credentials for mutual TLS if needed. use_client_cert = bool(util.strtobool(os.getenv("GOOGLE_API_USE_CLIENT_CERTIFICATE", "false"))) client_cert_source_func = None is_mtls = False if use_client_cert: if client_options.client_cert_source: is_mtls = True client_cert_source_func = client_options.client_cert_source else: is_mtls = mtls.has_default_client_cert_source() if is_mtls: client_cert_source_func = mtls.default_client_cert_source() else: client_cert_source_func = None # Figure out which api endpoint to use. if client_options.api_endpoint is not None: api_endpoint = client_options.api_endpoint else: use_mtls_env = os.getenv("GOOGLE_API_USE_MTLS_ENDPOINT", "auto") if use_mtls_env == "never": api_endpoint = self.DEFAULT_ENDPOINT elif use_mtls_env == "always": api_endpoint = self.DEFAULT_MTLS_ENDPOINT elif use_mtls_env == "auto": if is_mtls: api_endpoint = self.DEFAULT_MTLS_ENDPOINT else: api_endpoint = self.DEFAULT_ENDPOINT else: raise MutualTLSChannelError( "Unsupported GOOGLE_API_USE_MTLS_ENDPOINT value. Accepted " "values: never, auto, always" ) # Save or instantiate the transport. # Ordinarily, we provide the transport, but allowing a custom transport # instance provides an extensibility point for unusual situations. if isinstance(transport, EntityTypesTransport): # transport is a EntityTypesTransport instance. if credentials or client_options.credentials_file: raise ValueError("When providing a transport instance, " "provide its credentials directly.") if client_options.scopes: raise ValueError( "When providing a transport instance, provide its scopes " "directly." ) self._transport = transport else: Transport = type(self).get_transport_class(transport) self._transport = Transport( credentials=credentials, credentials_file=client_options.credentials_file, host=api_endpoint, scopes=client_options.scopes, client_cert_source_for_mtls=client_cert_source_func, quota_project_id=client_options.quota_project_id, client_info=client_info, ) def list_entity_types(self, request: entity_type.ListEntityTypesRequest = None, *, parent: str = None, retry: retries.Retry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> pagers.ListEntityTypesPager: r"""Returns the list of all entity types in the specified agent. Args: request (google.cloud.dialogflowcx_v3beta1.types.ListEntityTypesRequest): The request object. The request message for [EntityTypes.ListEntityTypes][google.cloud.dialogflow.cx.v3beta1.EntityTypes.ListEntityTypes]. parent (str): Required. The agent to list all entity types for. Format: ``projects/<Project ID>/locations/<Location ID>/agents/<Agent ID>``. This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.cloud.dialogflowcx_v3beta1.services.entity_types.pagers.ListEntityTypesPager: The response message for [EntityTypes.ListEntityTypes][google.cloud.dialogflow.cx.v3beta1.EntityTypes.ListEntityTypes]. Iterating over this object will yield results and resolve additional pages automatically. """ # Create or coerce a protobuf request object. # Sanity check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([parent]) if request is not None and has_flattened_params: raise ValueError('If the `request` argument is set, then none of ' 'the individual field arguments should be set.') # Minor optimization to avoid making a copy if the user passes # in a entity_type.ListEntityTypesRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance(request, entity_type.ListEntityTypesRequest): request = entity_type.ListEntityTypesRequest(request) # If we have keyword arguments corresponding to fields on the # request, apply these. if parent is not None: request.parent = parent # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[self._transport.list_entity_types] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata(( ("parent", request.parent), )), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # This method is paged; wrap the response in a pager, which provides # an `__iter__` convenience method. response = pagers.ListEntityTypesPager( method=rpc, request=request, response=response, metadata=metadata, ) # Done; return the response. return response def get_entity_type(self, request: entity_type.GetEntityTypeRequest = None, *, name: str = None, retry: retries.Retry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> entity_type.EntityType: r"""Retrieves the specified entity type. Args: request (google.cloud.dialogflowcx_v3beta1.types.GetEntityTypeRequest): The request object. The request message for [EntityTypes.GetEntityType][google.cloud.dialogflow.cx.v3beta1.EntityTypes.GetEntityType]. name (str): Required. The name of the entity type. Format: ``projects/<Project ID>/locations/<Location ID>/agents/<Agent ID>/entityTypes/<Entity Type ID>``. This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.cloud.dialogflowcx_v3beta1.types.EntityType: Entities are extracted from user input and represent parameters that are meaningful to your application. For example, a date range, a proper name such as a geographic location or landmark, and so on. Entities represent actionable data for your application. When you define an entity, you can also include synonyms that all map to that entity. For example, "soft drink", "soda", "pop", and so on. There are three types of entities: - **System** - entities that are defined by the Dialogflow API for common data types such as date, time, currency, and so on. A system entity is represented by the EntityType type. - **Custom** - entities that are defined by you that represent actionable data that is meaningful to your application. For example, you could define a pizza.sauce entity for red or white pizza sauce, a pizza.cheese entity for the different types of cheese on a pizza, a pizza.topping entity for different toppings, and so on. A custom entity is represented by the EntityType type. - **User** - entities that are built for an individual user such as favorites, preferences, playlists, and so on. A user entity is represented by the [SessionEntityType][google.cloud.dialogflow.cx.v3beta1.SessionEntityType] type. For more information about entity types, see the [Dialogflow documentation](\ https://cloud.google.com/dialogflow/docs/entities-overview). """ # Create or coerce a protobuf request object. # Sanity check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([name]) if request is not None and has_flattened_params: raise ValueError('If the `request` argument is set, then none of ' 'the individual field arguments should be set.') # Minor optimization to avoid making a copy if the user passes # in a entity_type.GetEntityTypeRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance(request, entity_type.GetEntityTypeRequest): request = entity_type.GetEntityTypeRequest(request) # If we have keyword arguments corresponding to fields on the # request, apply these. if name is not None: request.name = name # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[self._transport.get_entity_type] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata(( ("name", request.name), )), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # Done; return the response. return response def create_entity_type(self, request: gcdc_entity_type.CreateEntityTypeRequest = None, *, parent: str = None, entity_type: gcdc_entity_type.EntityType = None, retry: retries.Retry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> gcdc_entity_type.EntityType: r"""Creates an entity type in the specified agent. Args: request (google.cloud.dialogflowcx_v3beta1.types.CreateEntityTypeRequest): The request object. The request message for [EntityTypes.CreateEntityType][google.cloud.dialogflow.cx.v3beta1.EntityTypes.CreateEntityType]. parent (str): Required. The agent to create a entity type for. Format: ``projects/<Project ID>/locations/<Location ID>/agents/<Agent ID>``. This corresponds to the ``parent`` field on the ``request`` instance; if ``request`` is provided, this should not be set. entity_type (google.cloud.dialogflowcx_v3beta1.types.EntityType): Required. The entity type to create. This corresponds to the ``entity_type`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.cloud.dialogflowcx_v3beta1.types.EntityType: Entities are extracted from user input and represent parameters that are meaningful to your application. For example, a date range, a proper name such as a geographic location or landmark, and so on. Entities represent actionable data for your application. When you define an entity, you can also include synonyms that all map to that entity. For example, "soft drink", "soda", "pop", and so on. There are three types of entities: - **System** - entities that are defined by the Dialogflow API for common data types such as date, time, currency, and so on. A system entity is represented by the EntityType type. - **Custom** - entities that are defined by you that represent actionable data that is meaningful to your application. For example, you could define a pizza.sauce entity for red or white pizza sauce, a pizza.cheese entity for the different types of cheese on a pizza, a pizza.topping entity for different toppings, and so on. A custom entity is represented by the EntityType type. - **User** - entities that are built for an individual user such as favorites, preferences, playlists, and so on. A user entity is represented by the [SessionEntityType][google.cloud.dialogflow.cx.v3beta1.SessionEntityType] type. For more information about entity types, see the [Dialogflow documentation](\ https://cloud.google.com/dialogflow/docs/entities-overview). """ # Create or coerce a protobuf request object. # Sanity check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([parent, entity_type]) if request is not None and has_flattened_params: raise ValueError('If the `request` argument is set, then none of ' 'the individual field arguments should be set.') # Minor optimization to avoid making a copy if the user passes # in a gcdc_entity_type.CreateEntityTypeRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance(request, gcdc_entity_type.CreateEntityTypeRequest): request = gcdc_entity_type.CreateEntityTypeRequest(request) # If we have keyword arguments corresponding to fields on the # request, apply these. if parent is not None: request.parent = parent if entity_type is not None: request.entity_type = entity_type # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[self._transport.create_entity_type] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata(( ("parent", request.parent), )), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # Done; return the response. return response def update_entity_type(self, request: gcdc_entity_type.UpdateEntityTypeRequest = None, *, entity_type: gcdc_entity_type.EntityType = None, update_mask: field_mask_pb2.FieldMask = None, retry: retries.Retry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> gcdc_entity_type.EntityType: r"""Updates the specified entity type. Note: You should always train a flow prior to sending it queries. See the `training documentation <https://cloud.google.com/dialogflow/cx/docs/concept/training>`__. Args: request (google.cloud.dialogflowcx_v3beta1.types.UpdateEntityTypeRequest): The request object. The request message for [EntityTypes.UpdateEntityType][google.cloud.dialogflow.cx.v3beta1.EntityTypes.UpdateEntityType]. entity_type (google.cloud.dialogflowcx_v3beta1.types.EntityType): Required. The entity type to update. This corresponds to the ``entity_type`` field on the ``request`` instance; if ``request`` is provided, this should not be set. update_mask (google.protobuf.field_mask_pb2.FieldMask): The mask to control which fields get updated. This corresponds to the ``update_mask`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. Returns: google.cloud.dialogflowcx_v3beta1.types.EntityType: Entities are extracted from user input and represent parameters that are meaningful to your application. For example, a date range, a proper name such as a geographic location or landmark, and so on. Entities represent actionable data for your application. When you define an entity, you can also include synonyms that all map to that entity. For example, "soft drink", "soda", "pop", and so on. There are three types of entities: - **System** - entities that are defined by the Dialogflow API for common data types such as date, time, currency, and so on. A system entity is represented by the EntityType type. - **Custom** - entities that are defined by you that represent actionable data that is meaningful to your application. For example, you could define a pizza.sauce entity for red or white pizza sauce, a pizza.cheese entity for the different types of cheese on a pizza, a pizza.topping entity for different toppings, and so on. A custom entity is represented by the EntityType type. - **User** - entities that are built for an individual user such as favorites, preferences, playlists, and so on. A user entity is represented by the [SessionEntityType][google.cloud.dialogflow.cx.v3beta1.SessionEntityType] type. For more information about entity types, see the [Dialogflow documentation](\ https://cloud.google.com/dialogflow/docs/entities-overview). """ # Create or coerce a protobuf request object. # Sanity check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([entity_type, update_mask]) if request is not None and has_flattened_params: raise ValueError('If the `request` argument is set, then none of ' 'the individual field arguments should be set.') # Minor optimization to avoid making a copy if the user passes # in a gcdc_entity_type.UpdateEntityTypeRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance(request, gcdc_entity_type.UpdateEntityTypeRequest): request = gcdc_entity_type.UpdateEntityTypeRequest(request) # If we have keyword arguments corresponding to fields on the # request, apply these. if entity_type is not None: request.entity_type = entity_type if update_mask is not None: request.update_mask = update_mask # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[self._transport.update_entity_type] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata(( ("entity_type.name", request.entity_type.name), )), ) # Send the request. response = rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) # Done; return the response. return response def delete_entity_type(self, request: entity_type.DeleteEntityTypeRequest = None, *, name: str = None, retry: retries.Retry = gapic_v1.method.DEFAULT, timeout: float = None, metadata: Sequence[Tuple[str, str]] = (), ) -> None: r"""Deletes the specified entity type. Note: You should always train a flow prior to sending it queries. See the `training documentation <https://cloud.google.com/dialogflow/cx/docs/concept/training>`__. Args: request (google.cloud.dialogflowcx_v3beta1.types.DeleteEntityTypeRequest): The request object. The request message for [EntityTypes.DeleteEntityType][google.cloud.dialogflow.cx.v3beta1.EntityTypes.DeleteEntityType]. name (str): Required. The name of the entity type to delete. Format: ``projects/<Project ID>/locations/<Location ID>/agents/<Agent ID>/entityTypes/<Entity Type ID>``. This corresponds to the ``name`` field on the ``request`` instance; if ``request`` is provided, this should not be set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ # Create or coerce a protobuf request object. # Sanity check: If we got a request object, we should *not* have # gotten any keyword arguments that map to the request. has_flattened_params = any([name]) if request is not None and has_flattened_params: raise ValueError('If the `request` argument is set, then none of ' 'the individual field arguments should be set.') # Minor optimization to avoid making a copy if the user passes # in a entity_type.DeleteEntityTypeRequest. # There's no risk of modifying the input as we've already verified # there are no flattened fields. if not isinstance(request, entity_type.DeleteEntityTypeRequest): request = entity_type.DeleteEntityTypeRequest(request) # If we have keyword arguments corresponding to fields on the # request, apply these. if name is not None: request.name = name # Wrap the RPC method; this adds retry and timeout information, # and friendly error handling. rpc = self._transport._wrapped_methods[self._transport.delete_entity_type] # Certain fields should be provided within the metadata header; # add these here. metadata = tuple(metadata) + ( gapic_v1.routing_header.to_grpc_metadata(( ("name", request.name), )), ) # Send the request. rpc( request, retry=retry, timeout=timeout, metadata=metadata, ) try: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo( gapic_version=pkg_resources.get_distribution( "google-cloud-dialogflowcx", ).version, ) except pkg_resources.DistributionNotFound: DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo() __all__ = ( "EntityTypesClient", )
apache-2.0
7,160,798,180,582,101,000
44.345433
180
0.601704
false
CloudVE/djcloudbridge
djcloudbridge/drf_routers.py
1
1991
from django.conf.urls import url from rest_framework import routers, viewsets from rest_framework_nested import routers as nested_routers class HybridRoutingMixin(object): """ Extends functionality of DefaultRouter adding possibility to register simple API views, not just Viewsets. Based on: http://stackoverflow.com/questions/18818179/routing-api-views-in-django-rest-framework http://stackoverflow.com/questions/18817988/using-django-rest-frameworks-browsable-api-with-apiviews """ def get_routes(self, viewset): """ Checks if the viewset is an instance of ViewSet, otherwise assumes it's a simple view and does not run original `get_routes` code. """ if issubclass(viewset, viewsets.ViewSetMixin): return super(HybridRoutingMixin, self).get_routes(viewset) return [] def get_urls(self): """ Append non-viewset views to the urls generated by the original `get_urls` method. """ # URLs for viewsets ret = super(HybridRoutingMixin, self).get_urls() # URLs for simple views for prefix, viewset, basename in self.registry: # Skip viewsets if issubclass(viewset, viewsets.ViewSetMixin): continue # URL regex regex = '{prefix}{trailing_slash}$'.format( prefix=prefix, trailing_slash=self.trailing_slash ) # The view name has to have suffix "-list" due to specifics # of the DefaultRouter implementation. ret.append(url(regex, viewset.as_view(), name='{0}-list'.format(basename))) return ret class HybridDefaultRouter(HybridRoutingMixin, routers.DefaultRouter): pass class HybridSimpleRouter(HybridRoutingMixin, routers.SimpleRouter): pass class HybridNestedRouter(HybridRoutingMixin, nested_routers.NestedSimpleRouter): pass
mit
7,499,356,688,837,228,000
29.630769
104
0.64892
false
brentjens/rm-synthesis
doc/source/conf.py
1
7970
# -*- coding: utf-8 -*- # # RM-Synthesis documentation build configuration file, created by # sphinx-quickstart on Mon Jan 30 10:43:10 2012. # # This file is execfile()d with the current directory set to its containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys, os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ----------------------------------------------------- # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be extensions # coming with Sphinx (named 'sphinx.ext.*') or your custom ones. extensions = ['sphinx.ext.autodoc', 'sphinx.ext.doctest', 'sphinx.ext.imgmath', 'sphinx.ext.viewcode'] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'RM-Synthesis' copyright = u'2012, M.A. Brentjens <[email protected]>' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '1' # The full version, including alpha/beta/rc tags. release = '0.9' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # -- Options for HTML output --------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'default' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. #html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'RM-Synthesisdoc' # -- Options for LaTeX output -------------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, author, documentclass [howto/manual]). latex_documents = [ ('index', 'RM-Synthesis.tex', u'RM-Synthesis Documentation', u'M.A. Brentjens \\textless{}[email protected]\\textgreater{}', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output -------------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'rm-synthesis', u'RM-Synthesis Documentation', [u'M.A. Brentjens <[email protected]>'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------------ # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'RM-Synthesis', u'RM-Synthesis Documentation', u'M.A. Brentjens <[email protected]>', 'RM-Synthesis', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote'
mit
-7,558,729,425,361,679,000
31.933884
102
0.705019
false
Disiok/poetry-seq2seq
model.py
1
22939
#! /usr/bin/env python #-*- coding:utf-8 -*- # standard import os from IPython import embed # framework import tensorflow as tf from tensorflow.contrib import seq2seq, rnn from tensorflow.python.layers.core import Dense os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' class Seq2SeqModel: """ Seq2Seq model based on tensorflow.contrib.seq2seq """ def __init__(self, config, mode): assert mode.lower() in ['train', 'predict'] self.config = config self.mode = mode.lower() self.cell_type = config['cell_type'] self.hidden_units = config['hidden_units'] self.bidirectional = config['bidirectional'] self.decoder_hidden_units = self.hidden_units * (2 if self.bidirectional else 1) self.depth = config['depth'] self.attention_type = config['attention_type'] self.embedding_size = config['embedding_size'] self.vocab_size = config['vocab_size'] self.num_encoder_symbols = config['num_encoder_symbols'] self.num_decoder_symbols = config['num_decoder_symbols'] self.use_residual = config['use_residual'] self.attn_input_feeding = config['attn_input_feeding'] self.use_dropout = config['use_dropout'] self.keep_prob = 1.0 - config['dropout_rate'] self.optimizer = config['optimizer'] self.learning_rate = config['learning_rate'] self.max_gradient_norm = config['max_gradient_norm'] self.global_step = tf.Variable(0, trainable=False, name='global_step') self.global_epoch_step = tf.Variable(0, trainable=False, name='global_epoch_step') self.increment_global_epoch_step_op = tf.assign(self.global_epoch_step, self.global_epoch_step + 1) self.dtype = tf.float16 if config['use_fp16'] else tf.float32 self.keep_prob_placeholder = tf.placeholder(self.dtype, shape=[], name='keep_prob') self.use_beamsearch_decode=False if self.mode == 'predict': self.beam_width = config['beam_width'] self.use_beamsearch_decode = True if self.beam_width > 1 else False self.max_decode_step = config['max_decode_step'] self.predict_mode = config['predict_mode'] elif self.mode == 'train': self.train_mode = config['train_mode'] self.sampling_probability = config['sampling_probability'] self.start_token = config['start_token'] self.end_token = config['end_token'] self.build_model() def build_model(self): print 'Building model...' # Build encoder and decoder networks self.init_placeholders() self.build_encoder() self.build_decoder() # Merge all the training summaries self.summary_op = tf.summary.merge_all() def init_placeholders(self): # TODO(sdsuo): Understand dropout self.keep_prob_placeholder = tf.placeholder(self.dtype, shape=[], name='keep_prob') # embedding_placeholder: [vocab_size, hidden_units] self.embedding_placeholder = tf.placeholder( name='embedding_placeholder', shape=[self.vocab_size, self.hidden_units], dtype=self.dtype ) self.embedding = tf.get_variable( name='embedding', shape=[self.vocab_size, self.hidden_units], trainable=False, ) self.assign_embedding_op = self.embedding.assign(self.embedding_placeholder) # encode_inputs: [batch_size, time_steps] self.encoder_inputs = tf.placeholder( name='encoder_inputs', shape=(None, None), dtype=tf.int32 ) # encoder_inputs_length: [batch_size] self.encoder_inputs_length = tf.placeholder( name='encoder_inputs_length', shape=(None,), dtype=tf.int32 ) # use dynamic batch_size based on input self.batch_size = tf.shape(self.encoder_inputs)[0] if self.mode == 'train': # decoder_inputs: [batch_size, max_time_steps] self.decoder_inputs = tf.placeholder( dtype=tf.int32, shape=(None, None), name='decoder_inputs' ) # decoder_inputs_length: [batch_size] self.decoder_inputs_length = tf.placeholder( dtype=tf.int32, shape=(None,), name='decoder_inputs_length' ) decoder_start_token = tf.ones( shape=[self.batch_size, 1], dtype=tf.int32 ) * self.start_token decoder_end_token = tf.ones( shape=[self.batch_size, 1], dtype=tf.int32 ) * self.end_token # decoder_inputs_train: [batch_size , max_time_steps + 1] # insert _GO symbol in front of each decoder input self.decoder_inputs_train = tf.concat([decoder_start_token, self.decoder_inputs], axis=1) # decoder_inputs_length_train: [batch_size] self.decoder_inputs_length_train = self.decoder_inputs_length + 1 # decoder_targets_train: [batch_size, max_time_steps + 1] # insert EOS symbol at the end of each decoder input self.decoder_targets_train = tf.concat([self.decoder_inputs, decoder_end_token], axis=1) def build_single_cell(self, hidden_units): if self.cell_type == 'gru': cell_type = rnn.GRUCell elif self.cell_type == 'lstm': cell_type = rnn.LSTMCell else: raise RuntimeError('Unknown cell type!') cell = cell_type(hidden_units) return cell def build_encoder_cell(self): multi_cell = rnn.MultiRNNCell([self.build_single_cell(self.hidden_units) for _ in range(self.depth)]) return multi_cell def build_encoder(self): print 'Building encoder...' with tf.variable_scope('encoder'): # embedded inputs: [batch_size, time_step, embedding_size] self.encoder_inputs_embedded = tf.nn.embedding_lookup( params=self.embedding, ids=self.encoder_inputs ) # TODO(sdsuo): Decide if we need a Dense input layer here if self.bidirectional: # Build encoder cell self.encoder_cell_fw = self.build_encoder_cell() self.encoder_cell_bw = self.build_encoder_cell() # Encode input sequences into context vectors # encoder_outputs: [batch_size, time_step, cell_output_size] # encoder_last_state: [batch_size, cell_output_size] self.encoder_outputs_fw_bw, self.encoder_last_state_fw_bw = tf.nn.bidirectional_dynamic_rnn( cell_fw=self.encoder_cell_fw, cell_bw=self.encoder_cell_bw, inputs=self.encoder_inputs_embedded, sequence_length=self.encoder_inputs_length, dtype=self.dtype, time_major=False ) self.encoder_outputs_fw, self.encoder_outputs_bw = self.encoder_outputs_fw_bw self.encoder_outputs = tf.concat([self.encoder_outputs_fw, self.encoder_outputs_bw], 2) self.encoder_last_state_fw, self.encoder_last_state_bw = self.encoder_last_state_fw_bw encoder_last_state_zipped = zip(self.encoder_last_state_fw, self.encoder_last_state_bw) encoder_last_state_list = [rnn.LSTMStateTuple(c=tf.concat([fw.c, bw.c], 1), h=tf.concat([fw.h, bw.h], 1)) for fw, bw in encoder_last_state_zipped] self.encoder_last_state = tuple(encoder_last_state_list) else: self.encoder_cell = self.build_encoder_cell() # Encode input sequences into context vectors # encoder_outputs: [batch_size, time_step, cell_output_size] # encoder_last_state: [batch_size, cell_output_size] self.encoder_outputs, self.encoder_last_state = tf.nn.dynamic_rnn( cell=self.encoder_cell, inputs=self.encoder_inputs_embedded, sequence_length=self.encoder_inputs_length, dtype=self.dtype, time_major=False ) def build_decoder_cell(self): # TODO(sdsuo): Read up and decide whether to use beam search self.attention_mechanism = seq2seq.BahdanauAttention( num_units=self.decoder_hidden_units, memory=self.encoder_outputs, memory_sequence_length=self.encoder_inputs_length ) self.decoder_cell_list = [ self.build_single_cell(self.decoder_hidden_units) for _ in range(self.depth) ] # NOTE(sdsuo): Not sure what this does yet def attn_decoder_input_fn(inputs, attention): if not self.attn_input_feeding: return inputs # Essential when use_residual=True _input_layer = Dense(self.decoder_hidden_units, dtype=self.dtype, name='attn_input_feeding') return _input_layer(rnn.array_ops.concat([inputs, attention], -1)) # NOTE(sdsuo): Attention mechanism is implemented only on the top decoder layer self.decoder_cell_list[-1] = seq2seq.AttentionWrapper( cell=self.decoder_cell_list[-1], attention_mechanism=self.attention_mechanism, attention_layer_size=self.decoder_hidden_units, cell_input_fn=attn_decoder_input_fn, initial_cell_state=self.encoder_last_state[-1], alignment_history=False, name='attention_wrapper' ) # NOTE(sdsuo): Not sure why this is necessary # To be compatible with AttentionWrapper, the encoder last state # of the top layer should be converted into the AttentionWrapperState form # We can easily do this by calling AttentionWrapper.zero_state # Also if beamsearch decoding is used, the batch_size argument in .zero_state # should be ${decoder_beam_width} times to the origianl batch_size if self.use_beamsearch_decode: batch_size = self.batch_size * self.beam_width else: batch_size = self.batch_size # NOTE(vera): important dimension here # embed() initial_state = [state for state in self.encoder_last_state] initial_state[-1] = self.decoder_cell_list[-1].zero_state( batch_size=batch_size, dtype=self.dtype ) decoder_initial_state = tuple(initial_state) return rnn.MultiRNNCell(self.decoder_cell_list), decoder_initial_state def build_train_decoder(self): self.decoder_inputs_embedded = tf.nn.embedding_lookup( params=self.embedding, ids=self.decoder_inputs_train ) if self.train_mode == 'ground_truth': training_helper = seq2seq.TrainingHelper( inputs=self.decoder_inputs_embedded, sequence_length=self.decoder_inputs_length_train, time_major=False, name='training_helper' ) elif self.train_mode == 'scheduled_sampling': training_helper = seq2seq.ScheduledEmbeddingTrainingHelper( inputs=self.decoder_inputs_embedded, sequence_length=self.decoder_inputs_length_train, embedding=lambda inputs: tf.nn.embedding_lookup(self.embedding, inputs), sampling_probability=self.sampling_probability, name='scheduled_embedding_training_helper' ) else: raise NotImplementedError('Train mode: {} is not yet implemented'.format(self.train_mode)) training_decoder = seq2seq.BasicDecoder( cell=self.decoder_cell, helper=training_helper, initial_state=self.decoder_initial_state, output_layer=self.output_layer ) max_decoder_length = tf.reduce_max(self.decoder_inputs_length_train) self.decoder_outputs_train, self.decoder_last_state_train, self.decoder_outputs_length_train = seq2seq.dynamic_decode( decoder=training_decoder, output_time_major=False, impute_finished=True, maximum_iterations=max_decoder_length ) # NOTE(sdsuo): Not sure why this is necessary self.decoder_logits_train = tf.identity(self.decoder_outputs_train.rnn_output) # Use argmax to extract decoder symbols to emit self.decoder_pred_train = tf.argmax( self.decoder_logits_train, axis=-1, name='decoder_pred_train' ) # masks: masking for valid and padded time steps, [batch_size, max_time_step + 1] masks = tf.sequence_mask( lengths=self.decoder_inputs_length_train, maxlen=max_decoder_length, dtype=self.dtype, name='masks' ) # Computes per word average cross-entropy over a batch # Internally calls 'nn_ops.sparse_softmax_cross_entropy_with_logits' by default self.loss = seq2seq.sequence_loss( logits=self.decoder_logits_train, targets=self.decoder_targets_train, weights=masks, average_across_timesteps=True, average_across_batch=True ) # Training summary for the current batch_loss tf.summary.scalar('loss', self.loss) # Contruct graphs for minimizing loss self.init_optimizer() def build_predict_decoder(self): # start_tokens: [batch_size,] start_tokens = tf.ones([self.batch_size,], tf.int32) * self.start_token end_token =self.end_token if not self.use_beamsearch_decode: # Helper to feed inputs for greedy decoding: use the argmax of the output if self.predict_mode == 'sample': print 'Building sample decoder...' decoding_helper = seq2seq.SampleEmbeddingHelper( start_tokens=start_tokens, end_token=end_token, embedding=lambda inputs: tf.nn.embedding_lookup(self.embedding, inputs) ) elif self.predict_mode == 'greedy': print 'Building greedy decoder...' decoding_helper = seq2seq.GreedyEmbeddingHelper( start_tokens=start_tokens, end_token=end_token, embedding=lambda inputs: tf.nn.embedding_lookup(self.embedding, inputs) ) else: raise NotImplementedError('Predict mode: {} is not yet implemented'.format(self.predict_mode)) inference_decoder = seq2seq.BasicDecoder( cell=self.decoder_cell, helper=decoding_helper, initial_state=self.decoder_initial_state, output_layer=self.output_layer ) else: raise NotImplementedError('Beamsearch decode is not yet implemented.') self.decoder_outputs_decode, self.decoder_last_state_decode,self.decoder_outputs_length_decode = seq2seq.dynamic_decode( decoder=inference_decoder, output_time_major=False, maximum_iterations=self.max_decode_step ) if not self.use_beamsearch_decode: self.decoder_pred_decode = tf.expand_dims(self.decoder_outputs_decode.sample_id, -1) else: raise NotImplementedError('{} mode is not recognized.'.format(self.mode)) def build_decoder(self): print 'Building decoder...' with tf.variable_scope('decoder'): # Building decoder_cell and decoder_initial_state self.decoder_cell, self.decoder_initial_state = self.build_decoder_cell() # Output projection layer to convert cell_outputs to logits self.output_layer = Dense(self.vocab_size, name='output_projection') if self.mode == 'train': self.build_train_decoder() elif self.mode == 'predict': self.build_predict_decoder() else: raise RuntimeError def init_optimizer(self): print("Setting optimizer..") # Gradients and SGD update operation for training the model trainable_params = tf.trainable_variables() if self.optimizer.lower() == 'adadelta': self.opt = tf.train.AdadeltaOptimizer(learning_rate=self.learning_rate) elif self.optimizer.lower() == 'adam': self.opt = tf.train.AdamOptimizer(learning_rate=self.learning_rate) elif self.optimizer.lower() == 'rmsprop': self.opt = tf.train.RMSPropOptimizer(learning_rate=self.learning_rate) else: self.opt = tf.train.GradientDescentOptimizer(learning_rate=self.learning_rate) # Compute gradients of loss w.r.t. all trainable variables gradients = tf.gradients(self.loss, trainable_params) # Clip gradients by a given maximum_gradient_norm clip_gradients, _ = tf.clip_by_global_norm(gradients, self.max_gradient_norm) # Update the model self.updates = self.opt.apply_gradients( zip(clip_gradients, trainable_params), global_step=self.global_step) def save(self, sess, saver, path, global_step=None): """ Args: sess: path: var_list: global_step: Returns: """ save_path = saver.save(sess, save_path=path, global_step=global_step) print 'Model saved at {}'.format(save_path) def restore(self, sess, saver, path): """ Args: sess: path: var_list: Returns: """ saver.restore(sess, save_path=path) print 'Model restored from {}'.format(path) def train(self, sess, encoder_inputs, encoder_inputs_length, decoder_inputs, decoder_inputs_length): """Run a train step of the model feeding the given inputs. Args: session: tensorflow session to use. encoder_inputs: a numpy int matrix of [batch_size, max_source_time_steps] to feed as encoder inputs encoder_inputs_length: a numpy int vector of [batch_size] to feed as sequence lengths for each element in the given batch decoder_inputs: a numpy int matrix of [batch_size, max_target_time_steps] to feed as decoder inputs decoder_inputs_length: a numpy int vector of [batch_size] to feed as sequence lengths for each element in the given batch Returns: A triple consisting of gradient norm (or None if we did not do backward), average perplexity, and the outputs. """ # Check if the model is in training mode if self.mode != 'train': raise ValueError('Train step can only be operated in train mode') input_feed = self.check_feeds(encoder_inputs, encoder_inputs_length, decoder_inputs, decoder_inputs_length, False) # TODO(sdsuo): Understand keep prob input_feed[self.keep_prob_placeholder.name] = self.keep_prob output_feed = [ self.updates, # Update Op that does optimization self.loss, # Loss for current batch self.summary_op # Training summary ] outputs = sess.run(output_feed, input_feed) return outputs[1], outputs[2] # loss, summary def predict(self, sess, encoder_inputs, encoder_inputs_length): input_feed = self.check_feeds(encoder_inputs, encoder_inputs_length, decoder_inputs=None, decoder_inputs_length=None, predict=True) # Input feeds for dropout input_feed[self.keep_prob_placeholder.name] = 1.0 output_feed = [self.decoder_pred_decode] outputs = sess.run(output_feed, input_feed) # GreedyDecoder: [batch_size, max_time_step] # BeamSearchDecoder: [batch_size, max_time_step, beam_width] return outputs[0] def init_vars(self, sess, embedding): sess.run([self.assign_embedding_op], feed_dict={ self.embedding_placeholder: embedding }) def check_feeds(self, encoder_inputs, encoder_inputs_length, decoder_inputs, decoder_inputs_length, predict): """ Args: encoder_inputs: a numpy int matrix of [batch_size, max_source_time_steps] to feed as encoder inputs encoder_inputs_length: a numpy int vector of [batch_size] to feed as sequence lengths for each element in the given batch decoder_inputs: a numpy int matrix of [batch_size, max_target_time_steps] to feed as decoder inputs decoder_inputs_length: a numpy int vector of [batch_size] to feed as sequence lengths for each element in the given batch predict: a scalar boolean that indicates predict mode Returns: A feed for the model that consists of encoder_inputs, encoder_inputs_length, decoder_inputs, decoder_inputs_length """ input_batch_size = encoder_inputs.shape[0] if input_batch_size != encoder_inputs_length.shape[0]: raise ValueError("Encoder inputs and their lengths must be equal in their " "batch_size, %d != %d" % (input_batch_size, encoder_inputs_length.shape[0])) if not predict: target_batch_size = decoder_inputs.shape[0] if target_batch_size != input_batch_size: raise ValueError("Encoder inputs and Decoder inputs must be equal in their " "batch_size, %d != %d" % (input_batch_size, target_batch_size)) if target_batch_size != decoder_inputs_length.shape[0]: raise ValueError("Decoder targets and their lengths must be equal in their " "batch_size, %d != %d" % (target_batch_size, decoder_inputs_length.shape[0])) input_feed = {} input_feed[self.encoder_inputs.name] = encoder_inputs input_feed[self.encoder_inputs_length.name] = encoder_inputs_length if not predict: input_feed[self.decoder_inputs.name] = decoder_inputs input_feed[self.decoder_inputs_length.name] = decoder_inputs_length return input_feed if __name__ == '__main__': model = Seq2SeqModel() embed()
mit
-8,786,661,998,776,761,000
39.314587
128
0.594097
false
examachine/pisi
tests/constantstests.py
1
2098
# Copyright (C) 2005, TUBITAK/UEKAE # # This program is free software; you can redistribute it and/or modify it under # the terms of the GNU General Public License as published by the Free # Software Foundation; either version 3 of the License, or (at your option) # any later version. # # Please read the COPYING file. # # Author: Eray Ozkural <[email protected]> import unittest import pisi.context as ctx class ContextTestCase(unittest.TestCase): def testConstness(self): const = ctx.const # test if we can get a const attribute? try: test = const.package_suffix self.assertNotEqual(test, "") except AttributeError: self.fail("Couldn't get const attribute") # test binding a new constant const.test = "test binding" # test re-binding (which is illegal) try: const.test = "test rebinding" # we shouldn't reach here self.fail("Rebinding a constant works. Something is wrong!") except: # we achived our goal with this error. infact, this is a # ConstError but we can't catch it directly here pass # test unbinding (which is also illegal) try: del const.test # we shouldn't reach here self.fail("Unbinding a constant works. Something is wrong!") except: # we achived our goal with this error. infact, this is a # ConstError but we can't catch it directly here pass def testConstValues(self): const = ctx.const constDict = { "actions_file": "actions.py", "setup_func": "setup", "metadata_xml": "metadata.xml" } for k in constDict.keys(): if hasattr(const, k): value = getattr(const, k) self.assertEqual(value, constDict[k]) else: self.fail("Constants does not have an attribute named %s" % k) suite = unittest.makeSuite(ContextTestCase)
gpl-3.0
-8,041,903,705,030,690,000
29.852941
79
0.585796
false
michalczaplinski/pitchfork
tests/test_pitchfork.py
1
1447
from pitchfork.pitchfork import search, Review import json import unittest class TestReview(unittest.TestCase): @classmethod def setUpClass(cls): cls.review = search('mogwai', 'come on') def test_review(self): self.assertIsInstance(self.review, Review) def test_review_album(self): self.assertEqual(self.review.album(), "Come On Die Young") def test_reviev_artist(self): self.assertEqual(self.review.artist(), 'Mogwai') def test_reviev_best_new_music(self): self.assertEqual(self.review.best_new_music(), True) def test_review_label(self): self.assertEqual(self.review.label(), 'Chemikal Underground') def test_review_year(self): self.assertEqual(self.review.year(), '1999/2014') def test_score(self): self.assertEqual(self.review.score(), 8.3) def test_editorial(self): self.assertTrue(self.review.editorial().startswith('Though few of their songs contain actual words')) def test_review_url(self): self.assertEqual(self.review.url, '/reviews/albums/19466-mogwai-come-on-die-young-deluxe-edition/') def test_review_to_json(self): input_dict = self.review._json_safe_dict() output_dict = json.loads(self.review.to_json()) for input_key in input_dict.keys(): self.assertEqual(output_dict[input_key], input_dict[input_key]) if __name__ == '__main__': unittest.main()
mit
450,088,298,809,097,200
30.456522
109
0.665515
false
korepwx/tfsnippet
tests/trainer/test_base_trainer.py
1
8778
import functools import numpy as np import pytest import tensorflow as tf from mock import Mock from tfsnippet.dataflows import DataFlow from tfsnippet.scaffold import TrainLoop, AnnealingVariable, EventKeys from tfsnippet.trainer import * from tfsnippet.utils import EventSource class BaseTrainerTestCase(tf.test.TestCase): def test_props(self): loop = Mock(valid_metric_name='valid_loss') t = BaseTrainer(loop) self.assertIs(loop, t.loop) self.assertIsInstance(t.events, EventSource) def test_add_and_remove_hooks(self): loop = Mock( valid_metric_name='valid_loss', print_logs=Mock(return_value=None, __repr__=lambda o: 'print_logs') ) df = Mock() eval1 = Evaluator(loop, 1., [], df) eval2 = Evaluator(loop, 2., [], df) anneal1 = AnnealingVariable('anneal1', 1., .5) anneal2 = AnnealingVariable('anneal2', 2., .5) # test add t = BaseTrainer(loop) t.log_after_steps(3) t.log_after_epochs(4) t.evaluate_after_steps( Mock(return_value=None, __repr__=lambda o: 'eval'), 5) t.evaluate_after_epochs( Mock(return_value=None, __repr__=lambda o: 'eval'), 6) t.anneal_after_steps( Mock(return_value=None, __repr__=lambda o: 'anneal'), 7) t.anneal_after_epochs( Mock(return_value=None, __repr__=lambda o: 'anneal'), 8) t.evaluate_after_steps(eval1, 9) t.evaluate_after_epochs(eval2, 10) t.anneal_after_steps(anneal1, 11) t.anneal_after_epochs(anneal2, 12) t.log_after(steps=13) t.log_after(epochs=14) t.evaluate_after( Mock(return_value=None, __repr__=lambda o: 'eval2'), steps=15 ) t.evaluate_after( Mock(return_value=None, __repr__=lambda o: 'eval2'), epochs=16 ) t.anneal_after( Mock(return_value=None, __repr__=lambda o: 'anneal2'), steps=17 ) t.anneal_after( Mock(return_value=None, __repr__=lambda o: 'anneal2'), epochs=18 ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.STEP_EVALUATION]), '[eval:step:5, {!r}:step:9, eval2:step:15]'.format(eval1.run) ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.STEP_ANNEALING]), '[anneal:step:7, {!r}:step:11, anneal2:step:17]'. format(anneal1.anneal) ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.STEP_LOGGING]), '[print_logs:step:3, print_logs:step:13]' ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.EPOCH_EVALUATION]), '[eval:epoch:6, {!r}:epoch:10, eval2:epoch:16]'.format(eval2.run) ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.EPOCH_ANNEALING]), '[anneal:epoch:8, {!r}:epoch:12, anneal2:epoch:18]'. format(anneal2.anneal) ) self.assertEqual( repr(t.events._event_handlers_map[EventKeys.EPOCH_LOGGING]), '[print_logs:epoch:4, print_logs:epoch:14]' ) # test remove t.remove_log_hooks() self.assertNotIn( EventKeys.STEP_LOGGING, t.events._event_handlers_map) self.assertNotIn( EventKeys.EPOCH_LOGGING, t.events._event_handlers_map) t.remove_validation_hooks() self.assertNotIn( EventKeys.STEP_EVALUATION, t.events._event_handlers_map) self.assertNotIn( EventKeys.EPOCH_EVALUATION, t.events._event_handlers_map) t.remove_annealing_hooks() self.assertNotIn( EventKeys.STEP_ANNEALING, t.events._event_handlers_map) self.assertNotIn( EventKeys.EPOCH_ANNEALING, t.events._event_handlers_map) # test error add func_list = [ t.log_after, functools.partial(t.evaluate_after, Mock()), functools.partial(t.anneal_after, Mock()), ] kwargs_list = [ {'steps': None, 'epochs': None}, {'steps': 1, 'epochs': 1} ] for func in func_list: for kwargs in kwargs_list: with pytest.raises( ValueError, match='One and only one of `epochs` and ' '`steps` should be specified'): func(**kwargs) def test_hook_freq(self): loop = Mock( valid_metric_name='valid_loss', print_logs=Mock(return_value=None, __repr__=lambda o: 'print_logs') ) t = BaseTrainer(loop) f = Mock() t.evaluate_after(f, steps=5) for i in range(1, 6): t.loop.step = i t.events.fire(EventKeys.STEP_EVALUATION, t) t.loop.step = 7 t.events.fire(EventKeys.STEP_EVALUATION, t) t.loop.step = 10 t.events.fire(EventKeys.STEP_EVALUATION, t) self.assertEqual(f.call_count, 2) def test_run(self): with self.test_session() as session: df = DataFlow.arrays([np.arange(6, dtype=np.float32)], batch_size=4) def log_event(m, trainer): logged_events.append((m, trainer)) logged_events = [] # test default loss weight and merged feed dict with TrainLoop([], max_epoch=2) as loop: t = BaseTrainer(loop) t._run_step = Mock(return_value=None) t._iter_steps = Mock(wraps=lambda: loop.iter_steps(df)) for key in [EventKeys.BEFORE_EPOCH, EventKeys.BEFORE_STEP, EventKeys.STEP_ANNEALING, EventKeys.STEP_EVALUATION, EventKeys.STEP_LOGGING, EventKeys.AFTER_STEP, EventKeys.EPOCH_ANNEALING, EventKeys.EPOCH_EVALUATION, EventKeys.EPOCH_LOGGING, EventKeys.AFTER_EPOCH]: t.events.on(key, functools.partial(log_event, key)) t.run() self.assertEqual(4, len(t._run_step.call_args_list)) for i, call_args in enumerate(t._run_step.call_args_list[:-2]): call_session, call_payload = call_args[0] self.assertIs(session, call_session) self.assertEqual(i + 1, call_payload[0]) self.assertIsInstance(call_payload[1], tuple) self.assertEqual(1, len(call_payload[1])) np.testing.assert_equal( np.arange(6, dtype=np.float32)[i * 4: (i + 1) * 4], call_payload[1][0] ) expected_logged_events = sum( [ [ (EventKeys.BEFORE_EPOCH, t), ] + sum([ [ (EventKeys.BEFORE_STEP, t), (EventKeys.STEP_EVALUATION, t), (EventKeys.STEP_ANNEALING, t), (EventKeys.STEP_LOGGING, t), (EventKeys.AFTER_STEP, t), ] for step in [0, 1] ], []) + [ (EventKeys.EPOCH_EVALUATION, t), (EventKeys.EPOCH_ANNEALING, t), (EventKeys.EPOCH_LOGGING, t), (EventKeys.AFTER_EPOCH, t) ] for epoch in [0, 1] ], [] ) self.assertListEqual(logged_events, expected_logged_events) # test re-entrant error with TrainLoop([], max_epoch=1) as loop: t = BaseTrainer(loop) t._run_step = Mock(return_value=None) t._iter_steps = Mock(wraps=lambda: loop.iter_steps(df)) def reentrant_error(trainer): self.assertIs(trainer, t) with pytest.raises( RuntimeError, match=r'`run\(\)` is not re-entrant'): t.run() reentrant_error = Mock(wraps=reentrant_error) t.events.on(EventKeys.AFTER_STEP, reentrant_error) t.run() self.assertTrue(reentrant_error.called)
mit
-1,548,108,472,786,013,200
37.5
80
0.503418
false
mancoast/CPythonPyc_test
fail/312_test_platform.py
1
7787
import sys import os import unittest import platform import subprocess from test import support class PlatformTest(unittest.TestCase): def test_architecture(self): res = platform.architecture() if hasattr(os, "symlink"): def test_architecture_via_symlink(self): # issue3762 def get(python): cmd = [python, '-c', 'import platform; print(platform.architecture())'] p = subprocess.Popen(cmd, stdout=subprocess.PIPE) return p.communicate() real = os.path.realpath(sys.executable) link = os.path.abspath(support.TESTFN) os.symlink(real, link) try: self.assertEqual(get(real), get(link)) finally: os.remove(link) def test_platform(self): for aliased in (False, True): for terse in (False, True): res = platform.platform(aliased, terse) def test_system(self): res = platform.system() def test_node(self): res = platform.node() def test_release(self): res = platform.release() def test_version(self): res = platform.version() def test_machine(self): res = platform.machine() def test_processor(self): res = platform.processor() def setUp(self): self.save_version = sys.version self.save_subversion = sys.subversion self.save_platform = sys.platform def tearDown(self): sys.version = self.save_version sys.subversion = self.save_subversion sys.platform = self.save_platform def test_sys_version(self): # Old test. for input, output in ( ('2.4.3 (#1, Jun 21 2006, 13:54:21) \n[GCC 3.3.4 (pre 3.3.5 20040809)]', ('CPython', '2.4.3', '', '', '1', 'Jun 21 2006 13:54:21', 'GCC 3.3.4 (pre 3.3.5 20040809)')), ('IronPython 1.0.60816 on .NET 2.0.50727.42', ('IronPython', '1.0.60816', '', '', '', '', '.NET 2.0.50727.42')), ('IronPython 1.0 (1.0.61005.1977) on .NET 2.0.50727.42', ('IronPython', '1.0.0', '', '', '', '', '.NET 2.0.50727.42')), ): # branch and revision are not "parsed", but fetched # from sys.subversion. Ignore them (name, version, branch, revision, buildno, builddate, compiler) \ = platform._sys_version(input) self.assertEqual( (name, version, '', '', buildno, builddate, compiler), output) # Tests for python_implementation(), python_version(), python_branch(), # python_revision(), python_build(), and python_compiler(). sys_versions = { ("2.6.1 (r261:67515, Dec 6 2008, 15:26:00) \n[GCC 4.0.1 (Apple Computer, Inc. build 5370)]", ('CPython', 'tags/r261', '67515'), self.save_platform) : ("CPython", "2.6.1", "tags/r261", "67515", ('r261:67515', 'Dec 6 2008 15:26:00'), 'GCC 4.0.1 (Apple Computer, Inc. build 5370)'), ("IronPython 2.0 (2.0.0.0) on .NET 2.0.50727.3053", None, "cli") : ("IronPython", "2.0.0", "", "", ("", ""), ".NET 2.0.50727.3053"), ("2.5 (trunk:6107, Mar 26 2009, 13:02:18) \n[Java HotSpot(TM) Client VM (\"Apple Computer, Inc.\")]", ('Jython', 'trunk', '6107'), "java1.5.0_16") : ("Jython", "2.5.0", "trunk", "6107", ('trunk:6107', 'Mar 26 2009'), "java1.5.0_16"), ("2.5.2 (63378, Mar 26 2009, 18:03:29)\n[PyPy 1.0.0]", ('PyPy', 'trunk', '63378'), self.save_platform) : ("PyPy", "2.5.2", "trunk", "63378", ('63378', 'Mar 26 2009'), "") } for (version_tag, subversion, sys_platform), info in \ sys_versions.items(): sys.version = version_tag if subversion is None: if hasattr(sys, "subversion"): del sys.subversion else: sys.subversion = subversion if sys_platform is not None: sys.platform = sys_platform self.assertEqual(platform.python_implementation(), info[0]) self.assertEqual(platform.python_version(), info[1]) self.assertEqual(platform.python_branch(), info[2]) self.assertEqual(platform.python_revision(), info[3]) self.assertEqual(platform.python_build(), info[4]) self.assertEqual(platform.python_compiler(), info[5]) def test_system_alias(self): res = platform.system_alias( platform.system(), platform.release(), platform.version(), ) def test_uname(self): res = platform.uname() self.assertTrue(any(res)) def test_java_ver(self): res = platform.java_ver() if sys.platform == 'java': self.assertTrue(all(res)) def test_win32_ver(self): res = platform.win32_ver() def test_mac_ver(self): res = platform.mac_ver() if platform.uname()[0] == 'Darwin': # We're on a MacOSX system, check that # the right version information is returned fd = os.popen('sw_vers', 'r') real_ver = None for ln in fd: if ln.startswith('ProductVersion:'): real_ver = ln.strip().split()[-1] break fd.close() self.assertFalse(real_ver is None) result_list = res[0].split('.') expect_list = real_ver.split('.') len_diff = len(result_list) - len(expect_list) # On Snow Leopard, sw_vers reports 10.6.0 as 10.6 if len_diff > 0: expect_list.extend(['0'] * len_diff) self.assertEquals(result_list, expect_list) # res[1] claims to contain # (version, dev_stage, non_release_version) # That information is no longer available self.assertEquals(res[1], ('', '', '')) if sys.byteorder == 'little': self.assertEquals(res[2], 'i386') else: self.assertEquals(res[2], 'PowerPC') def test_dist(self): res = platform.dist() def test_libc_ver(self): import os if os.path.isdir(sys.executable) and \ os.path.exists(sys.executable+'.exe'): # Cygwin horror executable = executable + '.exe' res = platform.libc_ver(sys.executable) def test_parse_release_file(self): for input, output in ( # Examples of release file contents: ('SuSE Linux 9.3 (x86-64)', ('SuSE Linux ', '9.3', 'x86-64')), ('SUSE LINUX 10.1 (X86-64)', ('SUSE LINUX ', '10.1', 'X86-64')), ('SUSE LINUX 10.1 (i586)', ('SUSE LINUX ', '10.1', 'i586')), ('Fedora Core release 5 (Bordeaux)', ('Fedora Core', '5', 'Bordeaux')), ('Red Hat Linux release 8.0 (Psyche)', ('Red Hat Linux', '8.0', 'Psyche')), ('Red Hat Linux release 9 (Shrike)', ('Red Hat Linux', '9', 'Shrike')), ('Red Hat Enterprise Linux release 4 (Nahant)', ('Red Hat Enterprise Linux', '4', 'Nahant')), ('CentOS release 4', ('CentOS', '4', None)), ('Rocks release 4.2.1 (Cydonia)', ('Rocks', '4.2.1', 'Cydonia')), ('', ('', '', '')), # If there's nothing there. ): self.assertEqual(platform._parse_release_file(input), output) def test_main(): support.run_unittest( PlatformTest ) if __name__ == '__main__': test_main()
gpl-3.0
-7,698,688,418,364,413,000
36.985366
113
0.516502
false
promediacorp/flask-blog
post.py
1
6890
import datetime import cgi from bson.objectid import ObjectId from helper_functions import * class Post: def __init__(self, default_config): self.collection = default_config['POSTS_COLLECTION'] self.response = {'error': None, 'data': None} self.debug_mode = default_config['DEBUG'] def get_posts(self, limit, skip, tag=None, search=None): self.response['error'] = None cond = {} if tag is not None: cond = {'tags': tag} elif search is not None: cond = {'$or': [ {'title': {'$regex': search, '$options': 'i'}}, {'body': {'$regex': search, '$options': 'i'}}, {'preview': {'$regex': search, '$options': 'i'}}]} try: cursor = self.collection.find(cond).sort( 'date', direction=-1).skip(skip).limit(limit) self.response['data'] = [] for post in cursor: if 'tags' not in post: post['tags'] = [] if 'comments' not in post: post['comments'] = [] if 'preview' not in post: post['preview'] = '' self.response['data'].append({'id': post['_id'], 'title': post['title'], 'body': post['body'], 'preview': post['preview'], 'date': post['date'], 'permalink': post['permalink'], 'tags': post['tags'], 'author': post['author'], 'comments': post['comments']}) except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Posts not found..' return self.response def get_post_by_permalink(self, permalink): self.response['error'] = None try: self.response['data'] = self.collection.find_one( {'permalink': permalink}) except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Post not found..' return self.response def get_post_by_id(self, post_id): self.response['error'] = None try: self.response['data'] = self.collection.find_one( {'_id': ObjectId(post_id)}) if self.response['data']: if 'tags' not in self.response['data']: self.response['data']['tags'] = '' else: self.response['data']['tags'] = ','.join( self.response['data']['tags']) if 'preview' not in self.response['data']: self.response['data']['preview'] = '' except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Post not found..' return self.response def get_total_count(self, tag=None, search=None): cond = {} if tag is not None: cond = {'tags': tag} elif search is not None: cond = {'$or': [ {'title': {'$regex': search, '$options': 'i'}}, {'body': {'$regex': search, '$options': 'i'}}, {'preview': {'$regex': search, '$options': 'i'}}]} return self.collection.find(cond).count() def get_tags(self): self.response['error'] = None try: self.response['data'] = self.collection.aggregate([ {'$unwind': '$tags'}, {'$group': {'_id': '$tags', 'count': {'$sum': 1}}}, {'$sort': {'count': -1}}, {'$limit': 10}, {'$project': {'title': '$_id', 'count': 1, '_id': 0}} ]) if self.response['data']['result']: self.response['data'] = self.response['data']['result'] else: self.response['data'] = [] except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Get tags error..' return self.response def create_new_post(self, post_data): self.response['error'] = None try: self.response['data'] = self.collection.insert(post_data) except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Adding post error..' return self.response def edit_post(self, post_id, post_data): self.response['error'] = None del post_data['date'] del post_data['permalink'] try: self.collection.update( {'_id': ObjectId(post_id)}, {"$set": post_data}, upsert=False) self.response['data'] = True except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Post update error..' return self.response def delete_post(self, post_id): self.response['error'] = None try: if self.get_post_by_id(post_id) and self.collection.remove({'_id': ObjectId(post_id)}): self.response['data'] = True else: self.response['data'] = False except Exception, e: self.print_debug_info(e, self.debug_mode) self.response['error'] = 'Deleting post error..' return self.response @staticmethod def validate_post_data(post_data): # permalink = random_string(12) whitespace = re.compile('\s') permalink = whitespace.sub("-", post_data['title']).lower() post_data['title'] = cgi.escape(post_data['title']) post_data['preview'] = cgi.escape(post_data['preview'], quote=True) post_data['body'] = cgi.escape(post_data['body'], quote=True) post_data['date'] = datetime.datetime.utcnow() post_data['permalink'] = permalink return post_data @staticmethod def print_debug_info(msg, show=False): if show: import sys import os error_color = '\033[32m' error_end = '\033[0m' error = {'type': sys.exc_info()[0].__name__, 'file': os.path.basename(sys.exc_info()[2].tb_frame.f_code.co_filename), 'line': sys.exc_info()[2].tb_lineno, 'details': str(msg)} print error_color print '\n\n---\nError type: %s in file: %s on line: %s\nError details: %s\n---\n\n'\ % (error['type'], error['file'], error['line'], error['details']) print error_end
mit
-2,047,660,337,586,939,400
36.650273
99
0.475907
false
andrewharvey/asgs-stylesheets
configure.py
1
4492
#!/usr/bin/env python # Copyright (c) 2011, Development Seed, Inc. # 2011, Andrew Harvey <[email protected]> # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are # met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in # the documentation and/or other materials provided with the # distribution. # * Neither the name of the Development Seed, Inc. nor the names of # its contributors may be used to endorse or promote products # derived from this software without specific prior written # permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS # IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED # TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A # PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER # OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. import json from sys import path from os.path import join import argparse ################################# ## argparse parser = argparse.ArgumentParser(description='Configure an MML file with datasource settings') parser.add_argument('--host', default='localhost') parser.add_argument('--port', default='5432') parser.add_argument('--dbname', default='abs') parser.add_argument('--user', default='abs') parser.add_argument('--password', default='abs') parser.add_argument('--srs', default='+proj=merc +a=6378137 +b=6378137 +lat_ts=0.0 +lon_0=0.0 +x_0=0.0 +y_0=0.0 +k=1.0 +units=m +nadgrids=@null +wktext +no_defs +over') parser.add_argument('--shapedir', default='./layers/') # Increase performance if you are only rendering a particular area by # specifying a bounding box to restrict queries. Format is "XMIN,YMIN,XMAX,YMAX" in the # same units as the database (probably spherical mercator meters). The # whole world is "-20037508.34,-20037508.34,20037508.34,20037508.34". # Leave blank to let Mapnik estimate. parser.add_argument('--extent', default='12570320.00,-5403474.50,17711958.00,-1636391.88') parser.add_argument('--mml', required=True) args = parser.parse_args() ################################# ## configure mml mml = join(path[0], args.mml + '/' + args.mml + '.mml') shoreline_300 = args.shapedir.rstrip('/') + '/shoreline_300.shp' processed_p = args.shapedir.rstrip('/') + '/processed_p.shp' with open(mml, 'r') as f: newf = json.loads(f.read()) f.closed with open(mml, 'w') as f: for layer in newf["Layer"]: if "Datasource" in layer: ds_type = layer["Datasource"].get("type") if ds_type and ds_type == "postgis": layer["Datasource"]["host"] = args.host layer["Datasource"]["port"] = args.port layer["Datasource"]["dbname"] = args.dbname layer["Datasource"]["user"] = args.user layer["Datasource"]["password"] = args.password layer["Datasource"]["extent"] = args.extent layer["srs"] = args.srs else: if layer["id"] == "shoreline_300": layer["Datasource"] = dict(); layer["Datasource"]["file"] = shoreline_300 layer["Datasource"]["type"] = 'shape' layer["geometry"] = 'polygon' layer["srs"] = '+proj=merc +a=6378137 +b=6378137 +lat_ts=0.0 +lon_0=0.0 +x_0=0.0 +y_0=0 +k=1.0 +units=m +nadgrids=@null +wktext +no_defs +over' elif layer["id"] == "processed_p": layer["Datasource"] = dict(); layer["Datasource"]["file"] = processed_p layer["Datasource"]["type"] = 'shape' layer["geometry"] = 'polygon' layer["srs"] = '+proj=merc +a=6378137 +b=6378137 +lat_ts=0.0 +lon_0=0.0 +x_0=0.0 +y_0=0 +k=1.0 +units=m +nadgrids=@null +wktext +no_defs +over' f.write(json.dumps(newf, indent=2)) f.closed
cc0-1.0
5,764,988,682,530,616,000
42.61165
168
0.672752
false
pedromagnus/kilink
kilink/kilink.py
1
8947
# Copyright 2011-2013 Facundo Batista # All Rigths Reserved """The server and main app for kilink.""" import json import logging import time from functools import update_wrapper from flask import ( Flask, jsonify, make_response, redirect, render_template, request, ) from flask.ext.assets import Environment from flask_babel import Babel from flask_babel import gettext as _ from sqlalchemy import create_engine import backend import loghelper from config import config, LANGUAGES from decorators import crossdomain from metrics import StatsdClient # set up flask app = Flask(__name__) app.config.from_object(__name__) app.config["STATIC_URL"] = 'static' app.config["STATIC_ROOT"] = 'static' app.config["PROPAGATE_EXCEPTIONS"] = False babel = Babel(app) # flask-assets assets = Environment(app) assets.init_app(app) # logger logger = logging.getLogger('kilink.kilink') # metrics metrics = StatsdClient("linkode") def nocache(f): """Decorator to make a page un-cacheable.""" def new_func(*args, **kwargs): """The new function.""" resp = make_response(f(*args, **kwargs)) resp.headers['Cache-Control'] = 'public, max-age=0' return resp return update_wrapper(new_func, f) def measure(metric_name): """Decorator generator to send metrics counting and with timing.""" def _decorator(oldf): """The decorator itself.""" def newf(*args, **kwargs): """The function to replace.""" tini = time.time() try: result = oldf(*args, **kwargs) except Exception, exc: name = "%s.error.%s" % (metric_name, exc.__class__.__name__) metrics.count(name, 1) raise else: tdelta = time.time() - tini metrics.count(metric_name + '.ok', 1) metrics.timing(metric_name, tdelta) return result # need to fix the name because it's used by flask newf.func_name = oldf.func_name return newf return _decorator # accesory pages @app.route('/about') @measure("about") def about(): """Show the about page.""" return render_template('_about.html') @app.route('/tools') @measure("tools") def tools(): """Show the tools page.""" return render_template('_tools.html') # views @app.route('/') @measure("index") def index(): """The base page.""" render_dict = { 'value': '', 'button_text': _('Create linkode'), 'kid_info': '', 'tree_info': json.dumps(False), } return render_template('_new.html', **render_dict) @app.route('/', methods=['POST']) @measure("server.create") def create(): """Create a kilink.""" content = request.form['content'] text_type = request.form['text_type'] logger.debug("Create start; type=%r size=%d", text_type, len(content)) if text_type[:6] == "auto: ": text_type = text_type[6:] klnk = kilinkbackend.create_kilink(content, text_type) url = "/%s" % (klnk.kid,) logger.debug("Create done; kid=%s", klnk.kid) return redirect(url, code=303) @app.route('/<kid>', methods=['POST']) @app.route('/<kid>/<parent>', methods=['POST']) @measure("server.update") def update(kid, parent=None): """Update a kilink.""" content = request.form['content'] text_type = request.form['text_type'] logger.debug("Update start; kid=%r parent=%r type=%r size=%d", kid, parent, text_type, len(content)) if parent is None: root = kilinkbackend.get_root_node(kid) parent = root.revno klnk = kilinkbackend.update_kilink(kid, parent, content, text_type) new_url = "/%s/%s" % (kid, klnk.revno) logger.debug("Update done; kid=%r revno=%r", klnk.kid, klnk.revno) return redirect(new_url, code=303) @app.route('/<kid>') @app.route('/<kid>/<revno>') @app.route('/l/<kid>') @app.route('/l/<kid>/<revno>') @nocache @measure("server.show") def show(kid, revno=None): """Show the kilink content""" # get the content logger.debug("Show start; kid=%r revno=%r", kid, revno) if revno is None: klnk = kilinkbackend.get_root_node(kid) revno = klnk.revno else: klnk = kilinkbackend.get_kilink(kid, revno) content = klnk.content text_type = klnk.text_type # get the tree tree, nodeq = build_tree(kid, revno) render_dict = { 'value': content, 'button_text': _('Save new version'), 'kid_info': "%s/%s" % (kid, revno), 'tree_info': json.dumps(tree) if tree != {} else False, 'current_revno': revno, 'text_type': text_type, } logger.debug("Show done; quantity=%d", nodeq) return render_template('_new.html', **render_dict) def build_tree(kid, revno): """Build the tree for a given kilink id.""" nodes = [] for treenode in kilinkbackend.get_kilink_tree(kid): url = "/%s/%s" % (kid, treenode.revno) parent = treenode.parent nodes.append({ 'order': treenode.order, 'parent': parent, 'revno': treenode.revno, 'url': url, 'timestamp': str(treenode.timestamp), 'selected': treenode.revno == revno, }) root = [n for n in nodes if n['parent'] is None][0] fringe = [root, ] while fringe: node = fringe.pop() children = [n for n in nodes if n['parent'] == node['revno']] node['contents'] = children fringe.extend(children) return root, len(nodes) #API @app.route('/api/1/linkodes/', methods=['POST']) @crossdomain(origin='*') @measure("api.create") def api_create(): """Create a kilink.""" content = request.form['content'] text_type = request.form.get('text_type', "") logger.debug("API create start; type=%r size=%d", text_type, len(content)) klnk = kilinkbackend.create_kilink(content, text_type) ret_json = jsonify(linkode_id=klnk.kid, revno=klnk.revno) response = make_response(ret_json) response.headers['Location'] = 'http://%s/%s/%s' % ( config["server_host"], klnk.kid, klnk.revno) logger.debug("API create done; kid=%s", klnk.kid) return response, 201 @app.route('/api/1/linkodes/<kid>', methods=['POST']) @crossdomain(origin='*') @measure("api.update") def api_update(kid): """Update a kilink.""" content = request.form['content'] parent = request.form['parent'] text_type = request.form['text_type'] logger.debug("API update start; kid=%r parent=%r type=%r size=%d", kid, parent, text_type, len(content)) try: klnk = kilinkbackend.update_kilink(kid, parent, content, text_type) except backend.KilinkNotFoundError: logger.debug("API update done; kid %r not found", kid) response = make_response() return response, 404 logger.debug("API update done; kid=%r revno=%r", klnk.kid, klnk.revno) ret_json = jsonify(revno=klnk.revno) response = make_response(ret_json) response.headers['Location'] = 'http://%s/%s/%s' % ( config["server_host"], klnk.kid, klnk.revno) return response, 201 @app.route('/api/1/linkodes/<kid>/<revno>', methods=['GET']) @app.route('/api/1/linkodes/<kid>', methods=['GET']) @crossdomain(origin='*') @measure("api.get") def api_get(kid, revno=None): """Get the kilink and revno content""" logger.debug("API get; kid=%r revno=%r", kid, revno) if revno is None: klnk = kilinkbackend.get_root_node(kid) revno = klnk.revno else: klnk = kilinkbackend.get_kilink(kid, revno) # get the tree tree, nodeq = build_tree(kid, revno) logger.debug("API get done; type=%r size=%d len_tree=%d", klnk.text_type, len(klnk.content), nodeq) ret_json = jsonify(content=klnk.content, text_type=klnk.text_type, tree=tree) return ret_json @app.errorhandler(backend.KilinkNotFoundError) def handle_not_found_error(error): """Return 404 on kilink not found""" if request.url_rule.endpoint.startswith('api_'): response = jsonify({'message': error.message}) else: response = render_template('_404.html') logger.debug(error.message) return response, 404 @babel.localeselector def get_locale(): """Return the best matched language supported.""" return request.accept_languages.best_match(LANGUAGES.keys()) if __name__ == "__main__": # load config config.load_file("configs/development.yaml") # log setup handlers = loghelper.setup_logging(config['log_directory'], verbose=True) for h in handlers: app.logger.addHandler(h) h.setLevel(logging.DEBUG) app.logger.setLevel(logging.DEBUG) # set up the backend engine = create_engine(config["db_engine"], echo=True) kilinkbackend = backend.KilinkBackend(engine) app.run(debug=True, host='0.0.0.0')
gpl-3.0
-927,307,090,625,272,600
27.676282
78
0.612831
false
locationlabs/confab
confab/options.py
1
4182
""" Options for managing Confab. """ from os import getcwd from os.path import basename from fabric.api import env, task from fabric.utils import _AttributeDict from difflib import unified_diff from magic import Magic from re import match def _should_render(mime_type): """ Return whether a template file of with a particular mime type should be rendered. Some files may need to be excluded from template rendering; such files will be copied verbatim. """ return next((True for pattern in ['text/', 'application/xml'] if match(pattern, mime_type)), False) def _is_empty(mime_type): """ Return whether a template file is an empty file. """ return mime_type == 'inode/x-empty' def _is_not_temporary(file_name): """ Return whether a file name does not represent a temporary file. When listing configuration files, we usually want temporary files to be ignored. """ return not file_name.endswith('~') def _is_not_internal(file_name): """ Return whether a file name does not represent internal usage. When listing configuration files, we want to omit internal files, especially if they are used as Jinja includes """ return not basename(file_name).startswith('_') def _filter_func(file_name): """ Return the default filter func, which excludes temporary and internal files. """ return _is_not_temporary(file_name) and _is_not_internal(file_name) def _get_mime_type(file_name): """ Return the mime type of a file. The mime_type will be used to determine if a configuration file is text. """ return Magic(mime=True).from_file(file_name) def _diff(a, b, fromfile=None, tofile=None): """ Return a diff using '---', '+++', and '@@' control lines. By default, uses unified_diff. """ return unified_diff(a, b, fromfile=fromfile, tofile=tofile) def _as_dict(module): """ Returns publicly names values in module's __dict__. """ try: return {k: v for k, v in module.__dict__.iteritems() if not k[0:1] == '_'} except AttributeError: return {} def _get_base_dir(): """ Returns the base directory for user's template and data directories. """ return env.environmentdef.directory or getcwd() # Options that control how confab runs. # # These are in opposition to options likely to changed # between different runs of confab, such as directories, # environments, roles, hosts, etc. options = _AttributeDict({ # Should yes be assumed for interactive prompts? 'assume_yes': False, # How to compute a file's mime_type? 'get_mime_type': _get_mime_type, # How to determine if a template should be rendered? 'should_render': _should_render, # How to determine if a template is an empty file? 'is_empty': _is_empty, # How do filter available templates within the jinja environment? 'filter_func': _filter_func, # How to determine diffs? 'diff': _diff, # How to get dictionary configuration from module data? 'module_as_dict': _as_dict, # Base directory for template and data directories. 'get_base_dir': _get_base_dir, # What is the name of the template directory? 'get_templates_dir': lambda: 'templates', # What is the name of the data directory? 'get_data_dir': lambda: 'data', # What is the name of the generated directory? 'get_generated_dir': lambda: 'generated', # What is the name of the remotes directory? 'get_remotes_dir': lambda: 'remotes', }) class Options(object): """ Context manager to temporarily set options. """ def __init__(self, **kwargs): self.kwargs = kwargs self.previous = {} def __enter__(self): for (k, v) in self.kwargs.iteritems(): self.previous[k] = options.get(k) options[k] = v return self def __exit__(self, exc_type, value, traceback): for k in self.kwargs.keys(): options[k] = self.previous[k] @task def assume_yes(): """ Set the option to ``assume_yes`` in other tasks. """ options.assume_yes = True
apache-2.0
-6,501,699,333,116,751,000
24.345455
96
0.646341
false
jpartogi/django-job-board
job_board/urls.py
1
1230
from django.conf.urls.defaults import * from django.contrib.sitemaps import FlatPageSitemap, GenericSitemap from job_board.models import Job from job_board.feeds import JobFeed from job_board.forms import JobForm from job_board.views import JobFormPreview, job_list, job_detail, job_search feeds = { 'jobs': JobFeed, } info_dict = { 'queryset': Job.objects.filter(), 'date_field': 'posted' } sitemaps = { 'flatpages': FlatPageSitemap, 'jobs': GenericSitemap(info_dict, priority=0.6), } urlpatterns = patterns('', url(r'^feed/(?P<url>.*)/$', 'django.contrib.syndication.views.feed', {'feed_dict': feeds}, name='job-feeds'), url(r'^sitemap.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': sitemaps}, name='job-sitemap'), url(r'^new/$', JobFormPreview(JobForm), name='job-form'), url(r'^(?P<object_id>\d+)/(?P<slug>[\w-]+)/$', job_detail, name='job-detail'), url(r'^wmd/', include('wmd.urls')), url(r'^search/$', job_search, name='job-search'), url(r'^$', job_list, name='job-list'), # This must be last after everything else has been evaluated )
bsd-3-clause
4,233,101,760,313,957,000
23.137255
86
0.604065
false
coco-project/coco
coco/core/auth/authentication_backends.py
1
3945
from coco.contract.errors import AuthenticationError, ConnectionError, \ UserNotFoundError from coco.core.helpers import get_internal_ldap_connected, get_user_backend_connected from coco.core.models import BackendGroup, BackendUser, \ CollaborationGroup from django.contrib.auth.models import User from django.core.exceptions import PermissionDenied import logging logger = logging.getLogger(__name__) class BackendProxyAuthentication(object): """ Class used to authenticate with the user backends. more info: https://docs.djangoproject.com/en/1.8/topics/auth/default/#django.contrib.auth.authenticate """ def authenticate(self, username=None, password=None): """ :inherit. """ # check if the user already exists in our system # if so, use the defined backend_pk for validating the credentials on the backend # if its a Django only user, disallow the login user = None if User.objects.filter(username=username).exists(): user = User.objects.get(username=username) if hasattr(user, 'backend_user'): username = user.backend_user.backend_pk else: return None # not allowed, Django only user try: internal_ldap = get_internal_ldap_connected() user_backend = get_user_backend_connected() user_backend.auth_user(username, password) if user is not None: # existing user if not user.check_password(password): user.set_password(password) # XXX: not needed. should we leave it empty? internal_ldap.set_user_password(username, password) user.save() else: # new user uid = BackendUser.generate_internal_uid() group = self.create_user_groups(username, uid) user = self.create_users(username, password, uid, group.backend_group) group.add_user(user.backend_user) if user.is_active: return user else: return None except AuthenticationError: raise PermissionDenied except UserNotFoundError: if user is not None: # exists locally but not on backend user.delete() except ConnectionError as ex: logger.exception(ex) return None finally: try: internal_ldap.disconnect() user_backend.disconnect() except: pass def create_user_groups(self, name, gid): """ Create the groups for the logging-in user. :param name: The name of the group to create. :param gid: The group's ID (on the backend). """ collaboration_group = CollaborationGroup( name=name, is_single_user_group=True ) collaboration_group.save() backend_group = BackendGroup( django_group=collaboration_group, backend_id=gid, backend_pk=name ) backend_group.save() return collaboration_group def create_users(self, username, password, uid, primary_group): """ Create the Django users for the logging-in user. :param username: The user's username. :param primary_group: The user's primary group. """ user = User(username=username, password=password) user.save() backend_user = BackendUser( django_user=user, backend_id=uid, backend_pk=username, primary_group=primary_group ) backend_user.save() return user def get_user(self, user_id): """ :inherit. """ try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
bsd-3-clause
-1,043,595,838,817,111,200
33.008621
106
0.586565
false
ADL175/data-structures
src/graph.py
1
7286
"""Creates a Graph data structure, featuring graph traversal and two shortest path algorithms.""" import timeit import random class Graph(object): """Define the Graph class structure.""" def __init__(self): """Make an empty dictionary.""" self.graph_dict = {} def add_node(self, value): """Check if node of given value exists in dictionary.If not, add it with an empty list.""" try: self.graph_dict[value] except KeyError: self.graph_dict[value] = [] def add_edge(self, val1, val2, weight=0): """Ensure that nodes of val1 and val2 exist (creating them if they don't.Then make an edge connecting val1 to val2.""" if [val1, val2, weight] not in self.edges(): self.add_node(val1) self.add_node(val2) self.graph_dict[val1].append([val2, weight]) else: raise KeyError("Edge already exists.") def nodes(self): """Return a list of all keys in dictionary.""" return list(self.graph_dict.keys()) def edges(self): """Return a list of all edges in dictionary.""" to_return = [] for keys, values in self.graph_dict.items(): for i in values: to_return.append([keys, i[0], i[1]]) return to_return def del_node(self, val): """Delete a node from the graph, and from all edge pointers.""" try: del self.graph_dict[val] for keys, values in self.graph_dict.items(): for i in values: if i[0] == val: values.remove(i) except KeyError: raise KeyError("No such node exists.") def del_edge(self, val1, val2): """Delete an edge from graph.""" try: for node in self.graph_dict[val1]: if node[0] == val2: self.graph_dict[val1].remove(node) except KeyError: raise KeyError("No such node exists.") def has_node(self, val): """Check if graph has a given node in it.""" try: self.graph_dict[val] return True except KeyError: return False def neighbors(self, val): """Return all nodes connected to given node.""" try: return self.graph_dict[val] except KeyError: raise KeyError("No such node exists.") def adjacent(self, val1, val2): """Return True if edge exists, else return false.""" try: self.graph_dict[val2] return len(list(filter(lambda node: node[0] == val2, self.graph_dict[val1]))) > 0 except KeyError: raise KeyError("Value given not in graph.") def depth_first_traversal(self, val): """Return a list of all nodes connected to given start pointbased on a depth first algorithm.""" from stack import Stack seen = [] next_up = Stack() try: while True: if val not in seen: seen.append(val) for i in self.graph_dict[val][::-1]: next_up.push(i) if len(next_up) == 0: break val = next_up.pop()[0] return seen except KeyError: raise KeyError('Given value does not exist.') def breadth_first_traversal(self, val): """Return a list of all nodes connected to given start pointbased on a breadth first algorithm.""" from que_ import Queue seen = [] next_up = Queue() try: while True: if val not in seen: seen.append(val) for i in self.graph_dict[val]: next_up.enqueue(i) if next_up.size() == 0: break val = next_up.dequeue()[0] return seen except KeyError: raise KeyError('Given value does not exist.') def dijkstra(self, val1, val2): """An implementation of Dijkstra's shortest path algorithm. Makes use of a priority queue to find the shortest path between two nodes. Returns the distance of the node from the original, as well as the most optimal path. """ from priorityq import Priority_Q the_list = self.nodes() the_queue = Priority_Q() to_return = {} for i in the_list: if i == val1: the_queue.insert(i, self.graph_dict[i], 0) else: the_queue.insert(i, self.graph_dict[i]) while len(the_queue.heap.heap) > 0: current = the_queue.pop() for neighbor in current['neighbors']: alt = current['dist'] + neighbor[1] the_queue.decrease_priority(neighbor[0], alt, current['value']) to_return[current['value']] = current path = [] curr = to_return[val2] while True: path.append(curr['value']) if curr['prev']: curr = to_return[curr['prev']] else: return [to_return[val2]['dist'], path[::-1]] def bellman_ford(self, vertex_source, target): """An implementation the Bellman Ford shortest path algorithm. Makes use of a priority queue to find the shortest path between two nodes. Returns the distance of the node from the original, as well as the most optimal path. """ vertices = self.nodes() list_edges = self.edges() distance = {} predecessor = {} for vertex_v in vertices: distance[vertex_v] = float('inf') predecessor[vertex_v] = None distance[vertex_source] = 0 for i in range(len(vertices)): for ji in list_edges: if distance[ji[0]] + ji[2] < distance[ji[1]]: distance[ji[1]] = distance[ji[0]] + ji[2] predecessor[ji[1]] = ji[0] for i in list_edges: if distance[i[0]] + i[2] < distance[i[1]]: raise ValueError('Graph contains a negative-weight cycle') path = [] curr = target while True: path.append(curr) if predecessor[curr]: curr = predecessor[curr] else: return [distance[target], path[::-1]] def wrapper(func, *args, **kwargs): #pragma no cover """Create a value for a function with a specific arguement called to it.""" def wrapped(): return func(*args, **kwargs) return wrapped #code found at http://pythoncentral.io/time-a-python-function/ if __name__ == '__main__': #This block included for the requirement portion of the traversals assignment. time_test_graph = Graph() for i in range(100): time_test_graph.add_edge(i, i*2 + 1) time_test_graph.add_edge(i, i*2 + 2) wrapped1 = wrapper(time_test_graph.breadth_first_traversal, 0) wrapped2 = wrapper(time_test_graph.depth_first_traversal, 0) print("Breadth first: ", timeit.timeit(wrapped1, number=10000)) print("Depth first: ", timeit.timeit(wrapped2, number=10000))
mit
-7,696,677,436,234,390,000
34.198068
126
0.542136
false
Harmon758/discord.py
discord/client.py
1
48861
""" The MIT License (MIT) Copyright (c) 2015-present Rapptz Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ from __future__ import annotations import asyncio import logging import signal import sys import traceback from typing import Any, Generator, List, Optional, Sequence, TYPE_CHECKING, TypeVar, Union import aiohttp from .user import User from .invite import Invite from .template import Template from .widget import Widget from .guild import Guild from .channel import _channel_factory from .enums import ChannelType from .mentions import AllowedMentions from .errors import * from .enums import Status, VoiceRegion from .flags import ApplicationFlags from .gateway import * from .activity import BaseActivity, create_activity from .voice_client import VoiceClient from .http import HTTPClient from .state import ConnectionState from . import utils from .object import Object from .backoff import ExponentialBackoff from .webhook import Webhook from .iterators import GuildIterator from .appinfo import AppInfo from .ui.view import View from .stage_instance import StageInstance __all__ = ( 'Client', ) if TYPE_CHECKING: from .abc import SnowflakeTime log = logging.getLogger(__name__) def _cancel_tasks(loop): tasks = {t for t in asyncio.all_tasks(loop=loop) if not t.done()} if not tasks: return log.info('Cleaning up after %d tasks.', len(tasks)) for task in tasks: task.cancel() loop.run_until_complete(asyncio.gather(*tasks, return_exceptions=True)) log.info('All tasks finished cancelling.') for task in tasks: if task.cancelled(): continue if task.exception() is not None: loop.call_exception_handler({ 'message': 'Unhandled exception during Client.run shutdown.', 'exception': task.exception(), 'task': task }) def _cleanup_loop(loop): try: _cancel_tasks(loop) loop.run_until_complete(loop.shutdown_asyncgens()) finally: log.info('Closing the event loop.') loop.close() class Client: r"""Represents a client connection that connects to Discord. This class is used to interact with the Discord WebSocket and API. A number of options can be passed to the :class:`Client`. Parameters ----------- max_messages: Optional[:class:`int`] The maximum number of messages to store in the internal message cache. This defaults to ``1000``. Passing in ``None`` disables the message cache. .. versionchanged:: 1.3 Allow disabling the message cache and change the default size to ``1000``. loop: Optional[:class:`asyncio.AbstractEventLoop`] The :class:`asyncio.AbstractEventLoop` to use for asynchronous operations. Defaults to ``None``, in which case the default event loop is used via :func:`asyncio.get_event_loop()`. connector: :class:`aiohttp.BaseConnector` The connector to use for connection pooling. proxy: Optional[:class:`str`] Proxy URL. proxy_auth: Optional[:class:`aiohttp.BasicAuth`] An object that represents proxy HTTP Basic Authorization. shard_id: Optional[:class:`int`] Integer starting at ``0`` and less than :attr:`.shard_count`. shard_count: Optional[:class:`int`] The total number of shards. application_id: :class:`int` The client's application ID. intents: :class:`Intents` The intents that you want to enable for the session. This is a way of disabling and enabling certain gateway events from triggering and being sent. If not given, defaults to a regularly constructed :class:`Intents` class. .. versionadded:: 1.5 member_cache_flags: :class:`MemberCacheFlags` Allows for finer control over how the library caches members. If not given, defaults to cache as much as possible with the currently selected intents. .. versionadded:: 1.5 chunk_guilds_at_startup: :class:`bool` Indicates if :func:`.on_ready` should be delayed to chunk all guilds at start-up if necessary. This operation is incredibly slow for large amounts of guilds. The default is ``True`` if :attr:`Intents.members` is ``True``. .. versionadded:: 1.5 status: Optional[:class:`.Status`] A status to start your presence with upon logging on to Discord. activity: Optional[:class:`.BaseActivity`] An activity to start your presence with upon logging on to Discord. allowed_mentions: Optional[:class:`AllowedMentions`] Control how the client handles mentions by default on every message sent. .. versionadded:: 1.4 heartbeat_timeout: :class:`float` The maximum numbers of seconds before timing out and restarting the WebSocket in the case of not receiving a HEARTBEAT_ACK. Useful if processing the initial packets take too long to the point of disconnecting you. The default timeout is 60 seconds. guild_ready_timeout: :class:`float` The maximum number of seconds to wait for the GUILD_CREATE stream to end before preparing the member cache and firing READY. The default timeout is 2 seconds. .. versionadded:: 1.4 assume_unsync_clock: :class:`bool` Whether to assume the system clock is unsynced. This applies to the ratelimit handling code. If this is set to ``True``, the default, then the library uses the time to reset a rate limit bucket given by Discord. If this is ``False`` then your system clock is used to calculate how long to sleep for. If this is set to ``False`` it is recommended to sync your system clock to Google's NTP server. .. versionadded:: 1.3 Attributes ----------- ws The websocket gateway the client is currently connected to. Could be ``None``. loop: :class:`asyncio.AbstractEventLoop` The event loop that the client uses for asynchronous operations. """ def __init__(self, *, loop=None, **options): self.ws = None self.loop = asyncio.get_event_loop() if loop is None else loop self._listeners = {} self.shard_id = options.get('shard_id') self.shard_count = options.get('shard_count') connector = options.pop('connector', None) proxy = options.pop('proxy', None) proxy_auth = options.pop('proxy_auth', None) unsync_clock = options.pop('assume_unsync_clock', True) self.http = HTTPClient(connector, proxy=proxy, proxy_auth=proxy_auth, unsync_clock=unsync_clock, loop=self.loop) self._handlers = { 'ready': self._handle_ready } self._hooks = { 'before_identify': self._call_before_identify_hook } self._connection = self._get_state(**options) self._connection.shard_count = self.shard_count self._closed = False self._ready = asyncio.Event() self._connection._get_websocket = self._get_websocket self._connection._get_client = lambda: self if VoiceClient.warn_nacl: VoiceClient.warn_nacl = False log.warning("PyNaCl is not installed, voice will NOT be supported") # internals def _get_websocket(self, guild_id=None, *, shard_id=None): return self.ws def _get_state(self, **options): return ConnectionState(dispatch=self.dispatch, handlers=self._handlers, hooks=self._hooks, http=self.http, loop=self.loop, **options) def _handle_ready(self): self._ready.set() @property def latency(self): """:class:`float`: Measures latency between a HEARTBEAT and a HEARTBEAT_ACK in seconds. This could be referred to as the Discord WebSocket protocol latency. """ ws = self.ws return float('nan') if not ws else ws.latency def is_ws_ratelimited(self): """:class:`bool`: Whether the websocket is currently rate limited. This can be useful to know when deciding whether you should query members using HTTP or via the gateway. .. versionadded:: 1.6 """ if self.ws: return self.ws.is_ratelimited() return False @property def user(self): """Optional[:class:`.ClientUser`]: Represents the connected client. ``None`` if not logged in.""" return self._connection.user @property def guilds(self): """List[:class:`.Guild`]: The guilds that the connected client is a member of.""" return self._connection.guilds @property def emojis(self): """List[:class:`.Emoji`]: The emojis that the connected client has.""" return self._connection.emojis @property def cached_messages(self): """Sequence[:class:`.Message`]: Read-only list of messages the connected client has cached. .. versionadded:: 1.1 """ return utils.SequenceProxy(self._connection._messages or []) @property def private_channels(self): """List[:class:`.abc.PrivateChannel`]: The private channels that the connected client is participating on. .. note:: This returns only up to 128 most recent private channels due to an internal working on how Discord deals with private channels. """ return self._connection.private_channels @property def voice_clients(self): """List[:class:`.VoiceProtocol`]: Represents a list of voice connections. These are usually :class:`.VoiceClient` instances. """ return self._connection.voice_clients @property def application_id(self): """Optional[:class:`int`]: The client's application ID. If this is not passed via ``__init__`` then this is retrieved through the gateway when an event contains the data. Usually after :func:`~discord.on_connect` is called. """ return self._connection.application_id @property def application_flags(self) -> ApplicationFlags: """:class:`~discord.ApplicationFlags`: The client's application flags. .. versionadded: 2.0 """ return self._connection.application_flags # type: ignore def is_ready(self): """:class:`bool`: Specifies if the client's internal cache is ready for use.""" return self._ready.is_set() async def _run_event(self, coro, event_name, *args, **kwargs): try: await coro(*args, **kwargs) except asyncio.CancelledError: pass except Exception: try: await self.on_error(event_name, *args, **kwargs) except asyncio.CancelledError: pass def _schedule_event(self, coro, event_name, *args, **kwargs): wrapped = self._run_event(coro, event_name, *args, **kwargs) # Schedules the task return asyncio.create_task(wrapped, name=f'discord.py: {event_name}') def dispatch(self, event, *args, **kwargs): log.debug('Dispatching event %s', event) method = 'on_' + event listeners = self._listeners.get(event) if listeners: removed = [] for i, (future, condition) in enumerate(listeners): if future.cancelled(): removed.append(i) continue try: result = condition(*args) except Exception as exc: future.set_exception(exc) removed.append(i) else: if result: if len(args) == 0: future.set_result(None) elif len(args) == 1: future.set_result(args[0]) else: future.set_result(args) removed.append(i) if len(removed) == len(listeners): self._listeners.pop(event) else: for idx in reversed(removed): del listeners[idx] try: coro = getattr(self, method) except AttributeError: pass else: self._schedule_event(coro, method, *args, **kwargs) async def on_error(self, event_method, *args, **kwargs): """|coro| The default error handler provided by the client. By default this prints to :data:`sys.stderr` however it could be overridden to have a different implementation. Check :func:`~discord.on_error` for more details. """ print(f'Ignoring exception in {event_method}', file=sys.stderr) traceback.print_exc() # hooks async def _call_before_identify_hook(self, shard_id, *, initial=False): # This hook is an internal hook that actually calls the public one. # It allows the library to have its own hook without stepping on the # toes of those who need to override their own hook. await self.before_identify_hook(shard_id, initial=initial) async def before_identify_hook(self, shard_id, *, initial=False): """|coro| A hook that is called before IDENTIFYing a session. This is useful if you wish to have more control over the synchronization of multiple IDENTIFYing clients. The default implementation sleeps for 5 seconds. .. versionadded:: 1.4 Parameters ------------ shard_id: :class:`int` The shard ID that requested being IDENTIFY'd initial: :class:`bool` Whether this IDENTIFY is the first initial IDENTIFY. """ if not initial: await asyncio.sleep(5.0) # login state management async def login(self, token): """|coro| Logs in the client with the specified credentials. Parameters ----------- token: :class:`str` The authentication token. Do not prefix this token with anything as the library will do it for you. Raises ------ :exc:`.LoginFailure` The wrong credentials are passed. :exc:`.HTTPException` An unknown HTTP related error occurred, usually when it isn't 200 or the known incorrect credentials passing status code. """ log.info('logging in using static token') await self.http.static_login(token.strip()) async def connect(self, *, reconnect=True): """|coro| Creates a websocket connection and lets the websocket listen to messages from Discord. This is a loop that runs the entire event system and miscellaneous aspects of the library. Control is not resumed until the WebSocket connection is terminated. Parameters ----------- reconnect: :class:`bool` If we should attempt reconnecting, either due to internet failure or a specific failure on Discord's part. Certain disconnects that lead to bad state will not be handled (such as invalid sharding payloads or bad tokens). Raises ------- :exc:`.GatewayNotFound` If the gateway to connect to Discord is not found. Usually if this is thrown then there is a Discord API outage. :exc:`.ConnectionClosed` The websocket connection has been terminated. """ backoff = ExponentialBackoff() ws_params = { 'initial': True, 'shard_id': self.shard_id, } while not self.is_closed(): try: coro = DiscordWebSocket.from_client(self, **ws_params) self.ws = await asyncio.wait_for(coro, timeout=60.0) ws_params['initial'] = False while True: await self.ws.poll_event() except ReconnectWebSocket as e: log.info('Got a request to %s the websocket.', e.op) self.dispatch('disconnect') ws_params.update(sequence=self.ws.sequence, resume=e.resume, session=self.ws.session_id) continue except (OSError, HTTPException, GatewayNotFound, ConnectionClosed, aiohttp.ClientError, asyncio.TimeoutError) as exc: self.dispatch('disconnect') if not reconnect: await self.close() if isinstance(exc, ConnectionClosed) and exc.code == 1000: # clean close, don't re-raise this return raise if self.is_closed(): return # If we get connection reset by peer then try to RESUME if isinstance(exc, OSError) and exc.errno in (54, 10054): ws_params.update(sequence=self.ws.sequence, initial=False, resume=True, session=self.ws.session_id) continue # We should only get this when an unhandled close code happens, # such as a clean disconnect (1000) or a bad state (bad token, no sharding, etc) # sometimes, discord sends us 1000 for unknown reasons so we should reconnect # regardless and rely on is_closed instead if isinstance(exc, ConnectionClosed): if exc.code == 4014: raise PrivilegedIntentsRequired(exc.shard_id) from None if exc.code != 1000: await self.close() raise retry = backoff.delay() log.exception("Attempting a reconnect in %.2fs", retry) await asyncio.sleep(retry) # Always try to RESUME the connection # If the connection is not RESUME-able then the gateway will invalidate the session. # This is apparently what the official Discord client does. ws_params.update(sequence=self.ws.sequence, resume=True, session=self.ws.session_id) async def close(self): """|coro| Closes the connection to Discord. """ if self._closed: return self._closed = True for voice in self.voice_clients: try: await voice.disconnect() except Exception: # if an error happens during disconnects, disregard it. pass if self.ws is not None and self.ws.open: await self.ws.close(code=1000) await self.http.close() self._ready.clear() def clear(self): """Clears the internal state of the bot. After this, the bot can be considered "re-opened", i.e. :meth:`is_closed` and :meth:`is_ready` both return ``False`` along with the bot's internal cache cleared. """ self._closed = False self._ready.clear() self._connection.clear() self.http.recreate() async def start(self, token, *, reconnect=True): """|coro| A shorthand coroutine for :meth:`login` + :meth:`connect`. Raises ------- TypeError An unexpected keyword argument was received. """ await self.login(token) await self.connect(reconnect=reconnect) def run(self, *args, **kwargs): """A blocking call that abstracts away the event loop initialisation from you. If you want more control over the event loop then this function should not be used. Use :meth:`start` coroutine or :meth:`connect` + :meth:`login`. Roughly Equivalent to: :: try: loop.run_until_complete(start(*args, **kwargs)) except KeyboardInterrupt: loop.run_until_complete(close()) # cancel all tasks lingering finally: loop.close() .. warning:: This function must be the last function to call due to the fact that it is blocking. That means that registration of events or anything being called after this function call will not execute until it returns. """ loop = self.loop try: loop.add_signal_handler(signal.SIGINT, lambda: loop.stop()) loop.add_signal_handler(signal.SIGTERM, lambda: loop.stop()) except NotImplementedError: pass async def runner(): try: await self.start(*args, **kwargs) finally: if not self.is_closed(): await self.close() def stop_loop_on_completion(f): loop.stop() future = asyncio.ensure_future(runner(), loop=loop) future.add_done_callback(stop_loop_on_completion) try: loop.run_forever() except KeyboardInterrupt: log.info('Received signal to terminate bot and event loop.') finally: future.remove_done_callback(stop_loop_on_completion) log.info('Cleaning up tasks.') _cleanup_loop(loop) if not future.cancelled(): try: return future.result() except KeyboardInterrupt: # I am unsure why this gets raised here but suppress it anyway return None # properties def is_closed(self): """:class:`bool`: Indicates if the websocket connection is closed.""" return self._closed @property def activity(self): """Optional[:class:`.BaseActivity`]: The activity being used upon logging in. """ return create_activity(self._connection._activity) @activity.setter def activity(self, value): if value is None: self._connection._activity = None elif isinstance(value, BaseActivity): self._connection._activity = value.to_dict() else: raise TypeError('activity must derive from BaseActivity.') @property def allowed_mentions(self): """Optional[:class:`~discord.AllowedMentions`]: The allowed mention configuration. .. versionadded:: 1.4 """ return self._connection.allowed_mentions @allowed_mentions.setter def allowed_mentions(self, value): if value is None or isinstance(value, AllowedMentions): self._connection.allowed_mentions = value else: raise TypeError(f'allowed_mentions must be AllowedMentions not {value.__class__!r}') @property def intents(self): """:class:`~discord.Intents`: The intents configured for this connection. .. versionadded:: 1.5 """ return self._connection.intents # helpers/getters @property def users(self): """List[:class:`~discord.User`]: Returns a list of all the users the bot can see.""" return list(self._connection._users.values()) def get_channel(self, id): """Returns a channel with the given ID. Parameters ----------- id: :class:`int` The ID to search for. Returns -------- Optional[Union[:class:`.abc.GuildChannel`, :class:`.abc.PrivateChannel`]] The returned channel or ``None`` if not found. """ return self._connection.get_channel(id) def get_stage_instance(self, id) -> Optional[StageInstance]: """Returns a stage instance with the given stage channel ID. .. versionadded:: 2.0 Parameters ----------- id: :class:`int` The ID to search for. Returns -------- Optional[:class:`StageInstance`] The returns stage instance of ``None`` if not found. """ from .channel import StageChannel channel = self._connection.get_channel(id) if isinstance(channel, StageChannel): return channel.instance def get_guild(self, id): """Returns a guild with the given ID. Parameters ----------- id: :class:`int` The ID to search for. Returns -------- Optional[:class:`.Guild`] The guild or ``None`` if not found. """ return self._connection._get_guild(id) def get_user(self, id): """Returns a user with the given ID. Parameters ----------- id: :class:`int` The ID to search for. Returns -------- Optional[:class:`~discord.User`] The user or ``None`` if not found. """ return self._connection.get_user(id) def get_emoji(self, id): """Returns an emoji with the given ID. Parameters ----------- id: :class:`int` The ID to search for. Returns -------- Optional[:class:`.Emoji`] The custom emoji or ``None`` if not found. """ return self._connection.get_emoji(id) def get_all_channels(self): """A generator that retrieves every :class:`.abc.GuildChannel` the client can 'access'. This is equivalent to: :: for guild in client.guilds: for channel in guild.channels: yield channel .. note:: Just because you receive a :class:`.abc.GuildChannel` does not mean that you can communicate in said channel. :meth:`.abc.GuildChannel.permissions_for` should be used for that. Yields ------ :class:`.abc.GuildChannel` A channel the client can 'access'. """ for guild in self.guilds: yield from guild.channels def get_all_members(self): """Returns a generator with every :class:`.Member` the client can see. This is equivalent to: :: for guild in client.guilds: for member in guild.members: yield member Yields ------ :class:`.Member` A member the client can see. """ for guild in self.guilds: yield from guild.members # listeners/waiters async def wait_until_ready(self): """|coro| Waits until the client's internal cache is all ready. """ await self._ready.wait() def wait_for(self, event, *, check=None, timeout=None): """|coro| Waits for a WebSocket event to be dispatched. This could be used to wait for a user to reply to a message, or to react to a message, or to edit a message in a self-contained way. The ``timeout`` parameter is passed onto :func:`asyncio.wait_for`. By default, it does not timeout. Note that this does propagate the :exc:`asyncio.TimeoutError` for you in case of timeout and is provided for ease of use. In case the event returns multiple arguments, a :class:`tuple` containing those arguments is returned instead. Please check the :ref:`documentation <discord-api-events>` for a list of events and their parameters. This function returns the **first event that meets the requirements**. Examples --------- Waiting for a user reply: :: @client.event async def on_message(message): if message.content.startswith('$greet'): channel = message.channel await channel.send('Say hello!') def check(m): return m.content == 'hello' and m.channel == channel msg = await client.wait_for('message', check=check) await channel.send(f'Hello {msg.author}!') Waiting for a thumbs up reaction from the message author: :: @client.event async def on_message(message): if message.content.startswith('$thumb'): channel = message.channel await channel.send('Send me that \N{THUMBS UP SIGN} reaction, mate') def check(reaction, user): return user == message.author and str(reaction.emoji) == '\N{THUMBS UP SIGN}' try: reaction, user = await client.wait_for('reaction_add', timeout=60.0, check=check) except asyncio.TimeoutError: await channel.send('\N{THUMBS DOWN SIGN}') else: await channel.send('\N{THUMBS UP SIGN}') Parameters ------------ event: :class:`str` The event name, similar to the :ref:`event reference <discord-api-events>`, but without the ``on_`` prefix, to wait for. check: Optional[Callable[..., :class:`bool`]] A predicate to check what to wait for. The arguments must meet the parameters of the event being waited for. timeout: Optional[:class:`float`] The number of seconds to wait before timing out and raising :exc:`asyncio.TimeoutError`. Raises ------- asyncio.TimeoutError If a timeout is provided and it was reached. Returns -------- Any Returns no arguments, a single argument, or a :class:`tuple` of multiple arguments that mirrors the parameters passed in the :ref:`event reference <discord-api-events>`. """ future = self.loop.create_future() if check is None: def _check(*args): return True check = _check ev = event.lower() try: listeners = self._listeners[ev] except KeyError: listeners = [] self._listeners[ev] = listeners listeners.append((future, check)) return asyncio.wait_for(future, timeout) # event registration def event(self, coro): """A decorator that registers an event to listen to. You can find more info about the events on the :ref:`documentation below <discord-api-events>`. The events must be a :ref:`coroutine <coroutine>`, if not, :exc:`TypeError` is raised. Example --------- .. code-block:: python3 @client.event async def on_ready(): print('Ready!') Raises -------- TypeError The coroutine passed is not actually a coroutine. """ if not asyncio.iscoroutinefunction(coro): raise TypeError('event registered must be a coroutine function') setattr(self, coro.__name__, coro) log.debug('%s has successfully been registered as an event', coro.__name__) return coro async def change_presence(self, *, activity=None, status=None, afk=False): """|coro| Changes the client's presence. Example --------- .. code-block:: python3 game = discord.Game("with the API") await client.change_presence(status=discord.Status.idle, activity=game) Parameters ---------- activity: Optional[:class:`.BaseActivity`] The activity being done. ``None`` if no currently active activity is done. status: Optional[:class:`.Status`] Indicates what status to change to. If ``None``, then :attr:`.Status.online` is used. afk: Optional[:class:`bool`] Indicates if you are going AFK. This allows the discord client to know how to handle push notifications better for you in case you are actually idle and not lying. Raises ------ :exc:`.InvalidArgument` If the ``activity`` parameter is not the proper type. """ if status is None: status = 'online' status_enum = Status.online elif status is Status.offline: status = 'invisible' status_enum = Status.offline else: status_enum = status status = str(status) await self.ws.change_presence(activity=activity, status=status, afk=afk) for guild in self._connection.guilds: me = guild.me if me is None: continue if activity is not None: me.activities = (activity,) else: me.activities = () me.status = status_enum # Guild stuff def fetch_guilds(self, *, limit: int = 100, before: SnowflakeTime = None, after: SnowflakeTime = None) -> List[Guild]: """Retrieves an :class:`.AsyncIterator` that enables receiving your guilds. .. note:: Using this, you will only receive :attr:`.Guild.owner`, :attr:`.Guild.icon`, :attr:`.Guild.id`, and :attr:`.Guild.name` per :class:`.Guild`. .. note:: This method is an API call. For general usage, consider :attr:`guilds` instead. Examples --------- Usage :: async for guild in client.fetch_guilds(limit=150): print(guild.name) Flattening into a list :: guilds = await client.fetch_guilds(limit=150).flatten() # guilds is now a list of Guild... All parameters are optional. Parameters ----------- limit: Optional[:class:`int`] The number of guilds to retrieve. If ``None``, it retrieves every guild you have access to. Note, however, that this would make it a slow operation. Defaults to ``100``. before: Union[:class:`.abc.Snowflake`, :class:`datetime.datetime`] Retrieves guilds before this date or object. If a datetime is provided, it is recommended to use a UTC aware datetime. If the datetime is naive, it is assumed to be local time. after: Union[:class:`.abc.Snowflake`, :class:`datetime.datetime`] Retrieve guilds after this date or object. If a datetime is provided, it is recommended to use a UTC aware datetime. If the datetime is naive, it is assumed to be local time. Raises ------ :exc:`.HTTPException` Getting the guilds failed. Yields -------- :class:`.Guild` The guild with the guild data parsed. """ return GuildIterator(self, limit=limit, before=before, after=after) async def fetch_template(self, code): """|coro| Gets a :class:`.Template` from a discord.new URL or code. Parameters ----------- code: Union[:class:`.Template`, :class:`str`] The Discord Template Code or URL (must be a discord.new URL). Raises ------- :exc:`.NotFound` The template is invalid. :exc:`.HTTPException` Getting the template failed. Returns -------- :class:`.Template` The template from the URL/code. """ code = utils.resolve_template(code) data = await self.http.get_template(code) return Template(data=data, state=self._connection) # type: ignore async def fetch_guild(self, guild_id): """|coro| Retrieves a :class:`.Guild` from an ID. .. note:: Using this, you will **not** receive :attr:`.Guild.channels`, :attr:`.Guild.members`, :attr:`.Member.activity` and :attr:`.Member.voice` per :class:`.Member`. .. note:: This method is an API call. For general usage, consider :meth:`get_guild` instead. Parameters ----------- guild_id: :class:`int` The guild's ID to fetch from. Raises ------ :exc:`.Forbidden` You do not have access to the guild. :exc:`.HTTPException` Getting the guild failed. Returns -------- :class:`.Guild` The guild from the ID. """ data = await self.http.get_guild(guild_id) return Guild(data=data, state=self._connection) async def create_guild(self, name: str, region: Optional[VoiceRegion] = None, icon: Any = None, *, code: str = None): """|coro| Creates a :class:`.Guild`. Bot accounts in more than 10 guilds are not allowed to create guilds. Parameters ---------- name: :class:`str` The name of the guild. region: :class:`.VoiceRegion` The region for the voice communication server. Defaults to :attr:`.VoiceRegion.us_west`. icon: :class:`bytes` The :term:`py:bytes-like object` representing the icon. See :meth:`.ClientUser.edit` for more details on what is expected. code: Optional[:class:`str`] The code for a template to create the guild with. .. versionadded:: 1.4 Raises ------ :exc:`.HTTPException` Guild creation failed. :exc:`.InvalidArgument` Invalid icon image format given. Must be PNG or JPG. Returns ------- :class:`.Guild` The guild created. This is not the same guild that is added to cache. """ if icon is not None: icon = utils._bytes_to_base64_data(icon) region = region or VoiceRegion.us_west region_value = region.value if code: data = await self.http.create_from_template(code, name, region_value, icon) else: data = await self.http.create_guild(name, region_value, icon) return Guild(data=data, state=self._connection) async def fetch_stage_instance(self, channel_id: int) -> StageInstance: """|coro| Gets a :class:`StageInstance` for a stage channel id. .. versionadded:: 2.0 Parameters ----------- channel_id: :class:`int` The stage channel ID. Raises ------- :exc:`.NotFound` The stage instance or channel could not be found. :exc:`.HTTPException` Getting the stage instance failed. Returns -------- :class:`StageInstance` The stage instance from the stage channel ID. """ data = await self.http.get_stage_instance(channel_id) guild = self.get_guild(int(data['guild_id'])) return StageInstance(guild=guild, state=self._connection, data=data) # type: ignore # Invite management async def fetch_invite(self, url: Union[Invite, str], *, with_counts: bool = True, with_expiration: bool = True) -> Invite: """|coro| Gets an :class:`.Invite` from a discord.gg URL or ID. .. note:: If the invite is for a guild you have not joined, the guild and channel attributes of the returned :class:`.Invite` will be :class:`.PartialInviteGuild` and :class:`.PartialInviteChannel` respectively. Parameters ----------- url: Union[:class:`.Invite`, :class:`str`] The Discord invite ID or URL (must be a discord.gg URL). with_counts: :class:`bool` Whether to include count information in the invite. This fills the :attr:`.Invite.approximate_member_count` and :attr:`.Invite.approximate_presence_count` fields. with_expiration: :class:`bool` Whether to include the expiration date of the invite. This fills the :attr:`.Invite.expires_at` field. .. versionadded:: 2.0 Raises ------- :exc:`.NotFound` The invite has expired or is invalid. :exc:`.HTTPException` Getting the invite failed. Returns -------- :class:`.Invite` The invite from the URL/ID. """ invite_id = utils.resolve_invite(url) data = await self.http.get_invite(invite_id, with_counts=with_counts, with_expiration=with_expiration) return Invite.from_incomplete(state=self._connection, data=data) async def delete_invite(self, invite: Union[Invite, str]) -> None: """|coro| Revokes an :class:`.Invite`, URL, or ID to an invite. You must have the :attr:`~.Permissions.manage_channels` permission in the associated guild to do this. Parameters ---------- invite: Union[:class:`.Invite`, :class:`str`] The invite to revoke. Raises ------- :exc:`.Forbidden` You do not have permissions to revoke invites. :exc:`.NotFound` The invite is invalid or expired. :exc:`.HTTPException` Revoking the invite failed. """ invite_id = utils.resolve_invite(invite) await self.http.delete_invite(invite_id) # Miscellaneous stuff async def fetch_widget(self, guild_id): """|coro| Gets a :class:`.Widget` from a guild ID. .. note:: The guild must have the widget enabled to get this information. Parameters ----------- guild_id: :class:`int` The ID of the guild. Raises ------- :exc:`.Forbidden` The widget for this guild is disabled. :exc:`.HTTPException` Retrieving the widget failed. Returns -------- :class:`.Widget` The guild's widget. """ data = await self.http.get_widget(guild_id) return Widget(state=self._connection, data=data) async def application_info(self): """|coro| Retrieves the bot's application information. Raises ------- :exc:`.HTTPException` Retrieving the information failed somehow. Returns -------- :class:`.AppInfo` The bot's application information. """ data = await self.http.application_info() if 'rpc_origins' not in data: data['rpc_origins'] = None return AppInfo(self._connection, data) async def fetch_user(self, user_id): """|coro| Retrieves a :class:`~discord.User` based on their ID. You do not have to share any guilds with the user to get this information, however many operations do require that you do. .. note:: This method is an API call. If you have :attr:`discord.Intents.members` and member cache enabled, consider :meth:`get_user` instead. Parameters ----------- user_id: :class:`int` The user's ID to fetch from. Raises ------- :exc:`.NotFound` A user with this ID does not exist. :exc:`.HTTPException` Fetching the user failed. Returns -------- :class:`~discord.User` The user you requested. """ data = await self.http.get_user(user_id) return User(state=self._connection, data=data) async def fetch_channel(self, channel_id): """|coro| Retrieves a :class:`.abc.GuildChannel` or :class:`.abc.PrivateChannel` with the specified ID. .. note:: This method is an API call. For general usage, consider :meth:`get_channel` instead. .. versionadded:: 1.2 Raises ------- :exc:`.InvalidData` An unknown channel type was received from Discord. :exc:`.HTTPException` Retrieving the channel failed. :exc:`.NotFound` Invalid Channel ID. :exc:`.Forbidden` You do not have permission to fetch this channel. Returns -------- Union[:class:`.abc.GuildChannel`, :class:`.abc.PrivateChannel`] The channel from the ID. """ data = await self.http.get_channel(channel_id) factory, ch_type = _channel_factory(data['type']) if factory is None: raise InvalidData('Unknown channel type {type} for channel ID {id}.'.format_map(data)) if ch_type in (ChannelType.group, ChannelType.private): channel = factory(me=self.user, data=data, state=self._connection) else: guild_id = int(data['guild_id']) guild = self.get_guild(guild_id) or Object(id=guild_id) channel = factory(guild=guild, state=self._connection, data=data) return channel async def fetch_webhook(self, webhook_id): """|coro| Retrieves a :class:`.Webhook` with the specified ID. Raises -------- :exc:`.HTTPException` Retrieving the webhook failed. :exc:`.NotFound` Invalid webhook ID. :exc:`.Forbidden` You do not have permission to fetch this webhook. Returns --------- :class:`.Webhook` The webhook you requested. """ data = await self.http.get_webhook(webhook_id) return Webhook.from_state(data, state=self._connection) async def create_dm(self, user): """|coro| Creates a :class:`.DMChannel` with this user. This should be rarely called, as this is done transparently for most people. .. versionadded:: 2.0 Parameters ----------- user: :class:`~discord.abc.Snowflake` The user to create a DM with. Returns ------- :class:`.DMChannel` The channel that was created. """ state = self._connection found = state._get_private_channel_by_user(user.id) if found: return found data = await state.http.start_private_message(user.id) return state.add_dm_channel(data) def add_view(self, view: View, *, message_id: Optional[int] = None) -> None: """Registers a :class:`~discord.ui.View` for persistent listening. This method should be used for when a view is comprised of components that last longer than the lifecycle of the program. Parameters ------------ view: :class:`discord.ui.View` The view to register for dispatching. message_id: Optional[:class:`int`] The message ID that the view is attached to. This is currently used to refresh the view's state during message update events. If not given then message update events are not propagated for the view. Raises ------- TypeError A view was not passed. ValueError The view is not persistent. A persistent view has no timeout and all their components have an explicitly provided custom_id. """ if not isinstance(view, View): raise TypeError(f'expected an instance of View not {view.__class__!r}') if not view.is_persistent(): raise ValueError('View is not persistent. Items need to have a custom_id set and View must have no timeout') self._connection.store_view(view, message_id) @property def persistent_views(self) -> Sequence[View]: """Sequence[:class:`View`]: A sequence of persistent views added to the client.""" return self._connection.persistent_views
mit
-2,710,249,609,055,593,000
32.238776
144
0.579951
false
twoodford/audiovisualizer
build-movie.py
1
1391
import os import os.path import subprocess import sys from PIL import Image LISTF = "_list.txt" def get_dimensions(fpath): #print(fpath) return Image.open(fpath).size def run(folder, outfile, framerate=30, outres=(1920,1080)): jpglist = [os.path.join(folder, f) for f in os.listdir(folder) if f.startswith("frame_")] dimen = get_dimensions(jpglist[0]) ratio = float(outres[1])/outres[0] if dimen[0]*ratio < dimen[1]: crop = (dimen[0], int(dimen[0]*ratio)) else: crop = (int(dimen[1]/ratio), dimen[1]) with open(LISTF, "w") as ltxt: for f in jpglist: ltxt.write("file '"+f+"'\n") fsel_args = ["-f", "concat", "-i", LISTF] rs_str = "".join(("crop=", str(crop[0]), ":", str(crop[1]),":0:0,scale=",str(outres[0]),":",str(outres[1]))) enc_flags = ["-pix_fmt", "yuv420p", "-preset", "veryslow", "-crf", "18"] args_final = ["ffmpeg", "-r", str(framerate)] + fsel_args + ["-vf", rs_str] + enc_flags + [outfile] print(" ".join(args_final)) subprocess.call(args_final) os.remove(LISTF) if __name__=="__main__": jpglist = [os.path.join(sys.argv[1], f) for f in os.listdir(sys.argv[1]) if f.startswith("frame_")] dimen = get_dimensions(jpglist[0]) dimen = (dimen[0] if dimen[0]%2==0 else dimen[0]-1, dimen[1] if dimen[1]%2==0 else dimen[1]-1) run(sys.argv[1], sys.argv[2], outres=dimen)
apache-2.0
-3,590,627,063,179,411,500
37.638889
112
0.594536
false
dsweet04/rekall
rekall-lib/rekall_lib/registry.py
1
5206
# Rekall Memory Forensics # Copyright (C) 2011 # # Copyright 2013 Google Inc. All Rights Reserved. # # Michael Cohen <[email protected]> # # ****************************************************** # # * This program is free software; you can redistribute it and/or # * modify it under the terms of the GNU General Public License # * as published by the Free Software Foundation; either version 2 # * of the License, or (at your option) any later version. # * # * This program is distributed in the hope that it will be useful, # * but WITHOUT ANY WARRANTY; without even the implied warranty of # * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # * GNU General Public License for more details. # * # * You should have received a copy of the GNU General Public License # * along with this program; if not, write to the Free Software # * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ***************************************************** """ This module implements a class registry. We scan the memory_plugins directory for all python files and add those classes which should be registered into their own lookup tables. These are then ordered as required. The rest of Rekall Memory Forensics will then call onto the registered classes when needed. The MetaclassRegistry automatically adds any derived class to the base class. This means that we do not need to go through a special initializating step, as soon as a module is imported, the plugin is registered. """ __author__ = "Michael Cohen <[email protected]>" class classproperty(property): """A property that can be called on classes.""" def __get__(self, cls, owner): return self.fget(owner) def memoize(f): cache = {} def helper(*args): cached = cache.get(args, memoize) if cached is not memoize: return cached cached = f(*args) cache[args] = cached return cached return helper class UniqueObjectIdMetaclass(type): """Give each object a unique ID. unlike id() this number will not be reused when the objects are destroyed, hence it can be used to identify identical objects without keeping these around. """ ID = 0 def __call__(cls, *args, **kwargs): res = super(UniqueObjectIdMetaclass, cls).__call__(*args, **kwargs) res._object_id = UniqueObjectIdMetaclass.ID # pylint: disable=protected-access UniqueObjectIdMetaclass.ID += 1 return res class UniqueObjectIdMixin(object): __metaclass__ = UniqueObjectIdMetaclass class MetaclassRegistry(UniqueObjectIdMetaclass): """Automatic Plugin Registration through metaclasses.""" def __init__(cls, name, bases, env_dict): super(MetaclassRegistry, cls).__init__(name, bases, env_dict) cls._install_constructors(cls) # Attach the classes dict to the baseclass and have all derived classes # use the same one: for base in bases: try: cls.classes = base.classes cls.classes_by_name = base.classes_by_name cls.plugin_feature = base.plugin_feature cls.top_level_class = base.top_level_class break except AttributeError: cls.classes = {} cls.classes_by_name = {} cls.plugin_feature = cls.__name__ # Keep a reference to the top level class cls.top_level_class = cls # The following should not be registered as they are abstract. Classes # are abstract if the have the __abstract attribute (note this is not # inheritable so each abstract class must be explicitely marked). abstract_attribute = "_%s__abstract" % name if getattr(cls, abstract_attribute, None): return if not cls.__name__.startswith("Abstract"): if cls.__name__ in cls.classes: raise RuntimeError( "Multiple definitions for class %s (%s)" % ( cls, cls.classes[cls.__name__])) cls.classes[cls.__name__] = cls name = getattr(cls, "name", None) # We expect that classes by name will collide, which is why each # value is a list of classes with that name. cls.classes_by_name.setdefault(name, []).append(cls) try: if cls.top_level_class.include_plugins_as_attributes: setattr(cls.top_level_class, cls.__name__, cls) except AttributeError: pass # Allow the class itself to initialize itself. cls_initializer = getattr(cls, "_class_init", None) if cls_initializer: cls_initializer() @classmethod def _install_constructors(mcs, cls): def ByName(self, name): for impl in self.classes.values(): if getattr(impl, "name", None) == name: return impl cls.ImplementationByName = classmethod(ByName) def ByClass(self, name): return self.classes.get(name) cls.ImplementationByClass = classmethod(ByClass)
gpl-2.0
-1,171,829,265,467,452,700
33.939597
87
0.619478
false
AutorestCI/azure-sdk-for-python
azure-mgmt-network/azure/mgmt/network/v2017_11_01/models/available_providers_list_country.py
1
1381
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- from msrest.serialization import Model class AvailableProvidersListCountry(Model): """Country details. :param country_name: The country name. :type country_name: str :param providers: A list of Internet service providers. :type providers: list[str] :param states: List of available states in the country. :type states: list[~azure.mgmt.network.v2017_11_01.models.AvailableProvidersListState] """ _attribute_map = { 'country_name': {'key': 'countryName', 'type': 'str'}, 'providers': {'key': 'providers', 'type': '[str]'}, 'states': {'key': 'states', 'type': '[AvailableProvidersListState]'}, } def __init__(self, country_name=None, providers=None, states=None): super(AvailableProvidersListCountry, self).__init__() self.country_name = country_name self.providers = providers self.states = states
mit
-3,422,688,781,120,634,400
36.324324
77
0.604634
false
droark/bitcoin
test/functional/rpc_rawtransaction.py
1
24591
#!/usr/bin/env python3 # Copyright (c) 2014-2019 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. """Test the rawtransaction RPCs. Test the following RPCs: - createrawtransaction - signrawtransactionwithwallet - sendrawtransaction - decoderawtransaction - getrawtransaction """ from collections import OrderedDict from decimal import Decimal from io import BytesIO from test_framework.messages import CTransaction, ToHex from test_framework.test_framework import BitcoinTestFramework from test_framework.util import assert_equal, assert_raises_rpc_error, connect_nodes_bi, hex_str_to_bytes class multidict(dict): """Dictionary that allows duplicate keys. Constructed with a list of (key, value) tuples. When dumped by the json module, will output invalid json with repeated keys, eg: >>> json.dumps(multidict([(1,2),(1,2)]) '{"1": 2, "1": 2}' Used to test calls to rpc methods with repeated keys in the json object.""" def __init__(self, x): dict.__init__(self, x) self.x = x def items(self): return self.x # Create one-input, one-output, no-fee transaction: class RawTransactionsTest(BitcoinTestFramework): def set_test_params(self): self.setup_clean_chain = True self.num_nodes = 3 self.extra_args = [["-addresstype=legacy", "-txindex"], ["-addresstype=legacy", "-txindex"], ["-addresstype=legacy", "-txindex"]] def skip_test_if_missing_module(self): self.skip_if_no_wallet() def setup_network(self): super().setup_network() connect_nodes_bi(self.nodes, 0, 2) def run_test(self): self.log.info('prepare some coins for multiple *rawtransaction commands') self.nodes[2].generate(1) self.sync_all() self.nodes[0].generate(101) self.sync_all() self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(),1.5) self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(),1.0) self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(),5.0) self.sync_all() self.nodes[0].generate(5) self.sync_all() self.log.info('Test getrawtransaction on genesis block coinbase returns an error') block = self.nodes[0].getblock(self.nodes[0].getblockhash(0)) assert_raises_rpc_error(-5, "The genesis block coinbase is not considered an ordinary transaction", self.nodes[0].getrawtransaction, block['merkleroot']) self.log.info('Check parameter types and required parameters of createrawtransaction') # Test `createrawtransaction` required parameters assert_raises_rpc_error(-1, "createrawtransaction", self.nodes[0].createrawtransaction) assert_raises_rpc_error(-1, "createrawtransaction", self.nodes[0].createrawtransaction, []) # Test `createrawtransaction` invalid extra parameters assert_raises_rpc_error(-1, "createrawtransaction", self.nodes[0].createrawtransaction, [], {}, 0, False, 'foo') # Test `createrawtransaction` invalid `inputs` txid = '1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000' assert_raises_rpc_error(-3, "Expected type array", self.nodes[0].createrawtransaction, 'foo', {}) assert_raises_rpc_error(-1, "JSON value is not an object as expected", self.nodes[0].createrawtransaction, ['foo'], {}) assert_raises_rpc_error(-1, "JSON value is not a string as expected", self.nodes[0].createrawtransaction, [{}], {}) assert_raises_rpc_error(-8, "txid must be of length 64 (not 3, for 'foo')", self.nodes[0].createrawtransaction, [{'txid': 'foo'}], {}) assert_raises_rpc_error(-8, "txid must be hexadecimal string (not 'ZZZ7bb8b1697ea987f3b223ba7819250cae33efacb068d23dc24859824a77844')", self.nodes[0].createrawtransaction, [{'txid': 'ZZZ7bb8b1697ea987f3b223ba7819250cae33efacb068d23dc24859824a77844'}], {}) assert_raises_rpc_error(-8, "Invalid parameter, missing vout key", self.nodes[0].createrawtransaction, [{'txid': txid}], {}) assert_raises_rpc_error(-8, "Invalid parameter, missing vout key", self.nodes[0].createrawtransaction, [{'txid': txid, 'vout': 'foo'}], {}) assert_raises_rpc_error(-8, "Invalid parameter, vout must be positive", self.nodes[0].createrawtransaction, [{'txid': txid, 'vout': -1}], {}) assert_raises_rpc_error(-8, "Invalid parameter, sequence number is out of range", self.nodes[0].createrawtransaction, [{'txid': txid, 'vout': 0, 'sequence': -1}], {}) # Test `createrawtransaction` invalid `outputs` address = self.nodes[0].getnewaddress() address2 = self.nodes[0].getnewaddress() assert_raises_rpc_error(-1, "JSON value is not an array as expected", self.nodes[0].createrawtransaction, [], 'foo') self.nodes[0].createrawtransaction(inputs=[], outputs={}) # Should not throw for backwards compatibility self.nodes[0].createrawtransaction(inputs=[], outputs=[]) assert_raises_rpc_error(-8, "Data must be hexadecimal string", self.nodes[0].createrawtransaction, [], {'data': 'foo'}) assert_raises_rpc_error(-5, "Invalid Bitcoin address", self.nodes[0].createrawtransaction, [], {'foo': 0}) assert_raises_rpc_error(-3, "Invalid amount", self.nodes[0].createrawtransaction, [], {address: 'foo'}) assert_raises_rpc_error(-3, "Amount out of range", self.nodes[0].createrawtransaction, [], {address: -1}) assert_raises_rpc_error(-8, "Invalid parameter, duplicated address: %s" % address, self.nodes[0].createrawtransaction, [], multidict([(address, 1), (address, 1)])) assert_raises_rpc_error(-8, "Invalid parameter, duplicated address: %s" % address, self.nodes[0].createrawtransaction, [], [{address: 1}, {address: 1}]) assert_raises_rpc_error(-8, "Invalid parameter, duplicate key: data", self.nodes[0].createrawtransaction, [], [{"data": 'aa'}, {"data": "bb"}]) assert_raises_rpc_error(-8, "Invalid parameter, duplicate key: data", self.nodes[0].createrawtransaction, [], multidict([("data", 'aa'), ("data", "bb")])) assert_raises_rpc_error(-8, "Invalid parameter, key-value pair must contain exactly one key", self.nodes[0].createrawtransaction, [], [{'a': 1, 'b': 2}]) assert_raises_rpc_error(-8, "Invalid parameter, key-value pair not an object as expected", self.nodes[0].createrawtransaction, [], [['key-value pair1'], ['2']]) # Test `createrawtransaction` invalid `locktime` assert_raises_rpc_error(-3, "Expected type number", self.nodes[0].createrawtransaction, [], {}, 'foo') assert_raises_rpc_error(-8, "Invalid parameter, locktime out of range", self.nodes[0].createrawtransaction, [], {}, -1) assert_raises_rpc_error(-8, "Invalid parameter, locktime out of range", self.nodes[0].createrawtransaction, [], {}, 4294967296) # Test `createrawtransaction` invalid `replaceable` assert_raises_rpc_error(-3, "Expected type bool", self.nodes[0].createrawtransaction, [], {}, 0, 'foo') self.log.info('Check that createrawtransaction accepts an array and object as outputs') tx = CTransaction() # One output tx.deserialize(BytesIO(hex_str_to_bytes(self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs={address: 99})))) assert_equal(len(tx.vout), 1) assert_equal( tx.serialize().hex(), self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs=[{address: 99}]), ) # Two outputs tx.deserialize(BytesIO(hex_str_to_bytes(self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs=OrderedDict([(address, 99), (address2, 99)]))))) assert_equal(len(tx.vout), 2) assert_equal( tx.serialize().hex(), self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs=[{address: 99}, {address2: 99}]), ) # Multiple mixed outputs tx.deserialize(BytesIO(hex_str_to_bytes(self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs=multidict([(address, 99), (address2, 99), ('data', '99')]))))) assert_equal(len(tx.vout), 3) assert_equal( tx.serialize().hex(), self.nodes[2].createrawtransaction(inputs=[{'txid': txid, 'vout': 9}], outputs=[{address: 99}, {address2: 99}, {'data': '99'}]), ) for type in ["bech32", "p2sh-segwit", "legacy"]: addr = self.nodes[0].getnewaddress("", type) addrinfo = self.nodes[0].getaddressinfo(addr) pubkey = addrinfo["scriptPubKey"] self.log.info('sendrawtransaction with missing prevtx info (%s)' %(type)) # Test `signrawtransactionwithwallet` invalid `prevtxs` inputs = [ {'txid' : txid, 'vout' : 3, 'sequence' : 1000}] outputs = { self.nodes[0].getnewaddress() : 1 } rawtx = self.nodes[0].createrawtransaction(inputs, outputs) prevtx = dict(txid=txid, scriptPubKey=pubkey, vout=3, amount=1) succ = self.nodes[0].signrawtransactionwithwallet(rawtx, [prevtx]) assert succ["complete"] if type == "legacy": del prevtx["amount"] succ = self.nodes[0].signrawtransactionwithwallet(rawtx, [prevtx]) assert succ["complete"] if type != "legacy": assert_raises_rpc_error(-3, "Missing amount", self.nodes[0].signrawtransactionwithwallet, rawtx, [ { "txid": txid, "scriptPubKey": pubkey, "vout": 3, } ]) assert_raises_rpc_error(-3, "Missing vout", self.nodes[0].signrawtransactionwithwallet, rawtx, [ { "txid": txid, "scriptPubKey": pubkey, "amount": 1, } ]) assert_raises_rpc_error(-3, "Missing txid", self.nodes[0].signrawtransactionwithwallet, rawtx, [ { "scriptPubKey": pubkey, "vout": 3, "amount": 1, } ]) assert_raises_rpc_error(-3, "Missing scriptPubKey", self.nodes[0].signrawtransactionwithwallet, rawtx, [ { "txid": txid, "vout": 3, "amount": 1 } ]) ######################################### # sendrawtransaction with missing input # ######################################### self.log.info('sendrawtransaction with missing input') inputs = [ {'txid' : "1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000", 'vout' : 1}] #won't exists outputs = { self.nodes[0].getnewaddress() : 4.998 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) rawtx = self.nodes[2].signrawtransactionwithwallet(rawtx) # This will raise an exception since there are missing inputs assert_raises_rpc_error(-25, "Missing inputs", self.nodes[2].sendrawtransaction, rawtx['hex']) ##################################### # getrawtransaction with block hash # ##################################### # make a tx by sending then generate 2 blocks; block1 has the tx in it tx = self.nodes[2].sendtoaddress(self.nodes[1].getnewaddress(), 1) block1, block2 = self.nodes[2].generate(2) self.sync_all() # We should be able to get the raw transaction by providing the correct block gottx = self.nodes[0].getrawtransaction(tx, True, block1) assert_equal(gottx['txid'], tx) assert_equal(gottx['in_active_chain'], True) # We should not have the 'in_active_chain' flag when we don't provide a block gottx = self.nodes[0].getrawtransaction(tx, True) assert_equal(gottx['txid'], tx) assert 'in_active_chain' not in gottx # We should not get the tx if we provide an unrelated block assert_raises_rpc_error(-5, "No such transaction found", self.nodes[0].getrawtransaction, tx, True, block2) # An invalid block hash should raise the correct errors assert_raises_rpc_error(-1, "JSON value is not a string as expected", self.nodes[0].getrawtransaction, tx, True, True) assert_raises_rpc_error(-8, "parameter 3 must be of length 64 (not 6, for 'foobar')", self.nodes[0].getrawtransaction, tx, True, "foobar") assert_raises_rpc_error(-8, "parameter 3 must be of length 64 (not 8, for 'abcd1234')", self.nodes[0].getrawtransaction, tx, True, "abcd1234") assert_raises_rpc_error(-8, "parameter 3 must be hexadecimal string (not 'ZZZ0000000000000000000000000000000000000000000000000000000000000')", self.nodes[0].getrawtransaction, tx, True, "ZZZ0000000000000000000000000000000000000000000000000000000000000") assert_raises_rpc_error(-5, "Block hash not found", self.nodes[0].getrawtransaction, tx, True, "0000000000000000000000000000000000000000000000000000000000000000") # Undo the blocks and check in_active_chain self.nodes[0].invalidateblock(block1) gottx = self.nodes[0].getrawtransaction(txid=tx, verbose=True, blockhash=block1) assert_equal(gottx['in_active_chain'], False) self.nodes[0].reconsiderblock(block1) assert_equal(self.nodes[0].getbestblockhash(), block2) ######################### # RAW TX MULTISIG TESTS # ######################### # 2of2 test addr1 = self.nodes[2].getnewaddress() addr2 = self.nodes[2].getnewaddress() addr1Obj = self.nodes[2].getaddressinfo(addr1) addr2Obj = self.nodes[2].getaddressinfo(addr2) # Tests for createmultisig and addmultisigaddress assert_raises_rpc_error(-5, "Invalid public key", self.nodes[0].createmultisig, 1, ["01020304"]) self.nodes[0].createmultisig(2, [addr1Obj['pubkey'], addr2Obj['pubkey']]) # createmultisig can only take public keys assert_raises_rpc_error(-5, "Invalid public key", self.nodes[0].createmultisig, 2, [addr1Obj['pubkey'], addr1]) # addmultisigaddress can take both pubkeys and addresses so long as they are in the wallet, which is tested here. mSigObj = self.nodes[2].addmultisigaddress(2, [addr1Obj['pubkey'], addr1])['address'] #use balance deltas instead of absolute values bal = self.nodes[2].getbalance() # send 1.2 BTC to msig adr txId = self.nodes[0].sendtoaddress(mSigObj, 1.2) self.sync_all() self.nodes[0].generate(1) self.sync_all() assert_equal(self.nodes[2].getbalance(), bal+Decimal('1.20000000')) #node2 has both keys of the 2of2 ms addr., tx should affect the balance # 2of3 test from different nodes bal = self.nodes[2].getbalance() addr1 = self.nodes[1].getnewaddress() addr2 = self.nodes[2].getnewaddress() addr3 = self.nodes[2].getnewaddress() addr1Obj = self.nodes[1].getaddressinfo(addr1) addr2Obj = self.nodes[2].getaddressinfo(addr2) addr3Obj = self.nodes[2].getaddressinfo(addr3) mSigObj = self.nodes[2].addmultisigaddress(2, [addr1Obj['pubkey'], addr2Obj['pubkey'], addr3Obj['pubkey']])['address'] txId = self.nodes[0].sendtoaddress(mSigObj, 2.2) decTx = self.nodes[0].gettransaction(txId) rawTx = self.nodes[0].decoderawtransaction(decTx['hex']) self.sync_all() self.nodes[0].generate(1) self.sync_all() #THIS IS AN INCOMPLETE FEATURE #NODE2 HAS TWO OF THREE KEY AND THE FUNDS SHOULD BE SPENDABLE AND COUNT AT BALANCE CALCULATION assert_equal(self.nodes[2].getbalance(), bal) #for now, assume the funds of a 2of3 multisig tx are not marked as spendable txDetails = self.nodes[0].gettransaction(txId, True) rawTx = self.nodes[0].decoderawtransaction(txDetails['hex']) vout = False for outpoint in rawTx['vout']: if outpoint['value'] == Decimal('2.20000000'): vout = outpoint break bal = self.nodes[0].getbalance() inputs = [{ "txid" : txId, "vout" : vout['n'], "scriptPubKey" : vout['scriptPubKey']['hex'], "amount" : vout['value']}] outputs = { self.nodes[0].getnewaddress() : 2.19 } rawTx = self.nodes[2].createrawtransaction(inputs, outputs) rawTxPartialSigned = self.nodes[1].signrawtransactionwithwallet(rawTx, inputs) assert_equal(rawTxPartialSigned['complete'], False) #node1 only has one key, can't comp. sign the tx rawTxSigned = self.nodes[2].signrawtransactionwithwallet(rawTx, inputs) assert_equal(rawTxSigned['complete'], True) #node2 can sign the tx compl., own two of three keys self.nodes[2].sendrawtransaction(rawTxSigned['hex']) rawTx = self.nodes[0].decoderawtransaction(rawTxSigned['hex']) self.sync_all() self.nodes[0].generate(1) self.sync_all() assert_equal(self.nodes[0].getbalance(), bal+Decimal('50.00000000')+Decimal('2.19000000')) #block reward + tx # 2of2 test for combining transactions bal = self.nodes[2].getbalance() addr1 = self.nodes[1].getnewaddress() addr2 = self.nodes[2].getnewaddress() addr1Obj = self.nodes[1].getaddressinfo(addr1) addr2Obj = self.nodes[2].getaddressinfo(addr2) self.nodes[1].addmultisigaddress(2, [addr1Obj['pubkey'], addr2Obj['pubkey']])['address'] mSigObj = self.nodes[2].addmultisigaddress(2, [addr1Obj['pubkey'], addr2Obj['pubkey']])['address'] mSigObjValid = self.nodes[2].getaddressinfo(mSigObj) txId = self.nodes[0].sendtoaddress(mSigObj, 2.2) decTx = self.nodes[0].gettransaction(txId) rawTx2 = self.nodes[0].decoderawtransaction(decTx['hex']) self.sync_all() self.nodes[0].generate(1) self.sync_all() assert_equal(self.nodes[2].getbalance(), bal) # the funds of a 2of2 multisig tx should not be marked as spendable txDetails = self.nodes[0].gettransaction(txId, True) rawTx2 = self.nodes[0].decoderawtransaction(txDetails['hex']) vout = False for outpoint in rawTx2['vout']: if outpoint['value'] == Decimal('2.20000000'): vout = outpoint break bal = self.nodes[0].getbalance() inputs = [{ "txid" : txId, "vout" : vout['n'], "scriptPubKey" : vout['scriptPubKey']['hex'], "redeemScript" : mSigObjValid['hex'], "amount" : vout['value']}] outputs = { self.nodes[0].getnewaddress() : 2.19 } rawTx2 = self.nodes[2].createrawtransaction(inputs, outputs) rawTxPartialSigned1 = self.nodes[1].signrawtransactionwithwallet(rawTx2, inputs) self.log.debug(rawTxPartialSigned1) assert_equal(rawTxPartialSigned1['complete'], False) #node1 only has one key, can't comp. sign the tx rawTxPartialSigned2 = self.nodes[2].signrawtransactionwithwallet(rawTx2, inputs) self.log.debug(rawTxPartialSigned2) assert_equal(rawTxPartialSigned2['complete'], False) #node2 only has one key, can't comp. sign the tx rawTxComb = self.nodes[2].combinerawtransaction([rawTxPartialSigned1['hex'], rawTxPartialSigned2['hex']]) self.log.debug(rawTxComb) self.nodes[2].sendrawtransaction(rawTxComb) rawTx2 = self.nodes[0].decoderawtransaction(rawTxComb) self.sync_all() self.nodes[0].generate(1) self.sync_all() assert_equal(self.nodes[0].getbalance(), bal+Decimal('50.00000000')+Decimal('2.19000000')) #block reward + tx # decoderawtransaction tests # witness transaction encrawtx = "010000000001010000000000000072c1a6a246ae63f74f931e8365e15a089c68d61900000000000000000000ffffffff0100e1f50500000000000102616100000000" decrawtx = self.nodes[0].decoderawtransaction(encrawtx, True) # decode as witness transaction assert_equal(decrawtx['vout'][0]['value'], Decimal('1.00000000')) assert_raises_rpc_error(-22, 'TX decode failed', self.nodes[0].decoderawtransaction, encrawtx, False) # force decode as non-witness transaction # non-witness transaction encrawtx = "01000000010000000000000072c1a6a246ae63f74f931e8365e15a089c68d61900000000000000000000ffffffff0100e1f505000000000000000000" decrawtx = self.nodes[0].decoderawtransaction(encrawtx, False) # decode as non-witness transaction assert_equal(decrawtx['vout'][0]['value'], Decimal('1.00000000')) # getrawtransaction tests # 1. valid parameters - only supply txid txHash = rawTx["hash"] assert_equal(self.nodes[0].getrawtransaction(txHash), rawTxSigned['hex']) # 2. valid parameters - supply txid and 0 for non-verbose assert_equal(self.nodes[0].getrawtransaction(txHash, 0), rawTxSigned['hex']) # 3. valid parameters - supply txid and False for non-verbose assert_equal(self.nodes[0].getrawtransaction(txHash, False), rawTxSigned['hex']) # 4. valid parameters - supply txid and 1 for verbose. # We only check the "hex" field of the output so we don't need to update this test every time the output format changes. assert_equal(self.nodes[0].getrawtransaction(txHash, 1)["hex"], rawTxSigned['hex']) # 5. valid parameters - supply txid and True for non-verbose assert_equal(self.nodes[0].getrawtransaction(txHash, True)["hex"], rawTxSigned['hex']) # 6. invalid parameters - supply txid and string "Flase" assert_raises_rpc_error(-1, "not a boolean", self.nodes[0].getrawtransaction, txHash, "Flase") # 7. invalid parameters - supply txid and empty array assert_raises_rpc_error(-1, "not a boolean", self.nodes[0].getrawtransaction, txHash, []) # 8. invalid parameters - supply txid and empty dict assert_raises_rpc_error(-1, "not a boolean", self.nodes[0].getrawtransaction, txHash, {}) inputs = [ {'txid' : "1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000", 'vout' : 1, 'sequence' : 1000}] outputs = { self.nodes[0].getnewaddress() : 1 } rawtx = self.nodes[0].createrawtransaction(inputs, outputs) decrawtx= self.nodes[0].decoderawtransaction(rawtx) assert_equal(decrawtx['vin'][0]['sequence'], 1000) # 9. invalid parameters - sequence number out of range inputs = [ {'txid' : "1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000", 'vout' : 1, 'sequence' : -1}] outputs = { self.nodes[0].getnewaddress() : 1 } assert_raises_rpc_error(-8, 'Invalid parameter, sequence number is out of range', self.nodes[0].createrawtransaction, inputs, outputs) # 10. invalid parameters - sequence number out of range inputs = [ {'txid' : "1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000", 'vout' : 1, 'sequence' : 4294967296}] outputs = { self.nodes[0].getnewaddress() : 1 } assert_raises_rpc_error(-8, 'Invalid parameter, sequence number is out of range', self.nodes[0].createrawtransaction, inputs, outputs) inputs = [ {'txid' : "1d1d4e24ed99057e84c3f80fd8fbec79ed9e1acee37da269356ecea000000000", 'vout' : 1, 'sequence' : 4294967294}] outputs = { self.nodes[0].getnewaddress() : 1 } rawtx = self.nodes[0].createrawtransaction(inputs, outputs) decrawtx= self.nodes[0].decoderawtransaction(rawtx) assert_equal(decrawtx['vin'][0]['sequence'], 4294967294) #################################### # TRANSACTION VERSION NUMBER TESTS # #################################### # Test the minimum transaction version number that fits in a signed 32-bit integer. tx = CTransaction() tx.nVersion = -0x80000000 rawtx = ToHex(tx) decrawtx = self.nodes[0].decoderawtransaction(rawtx) assert_equal(decrawtx['version'], -0x80000000) # Test the maximum transaction version number that fits in a signed 32-bit integer. tx = CTransaction() tx.nVersion = 0x7fffffff rawtx = ToHex(tx) decrawtx = self.nodes[0].decoderawtransaction(rawtx) assert_equal(decrawtx['version'], 0x7fffffff) if __name__ == '__main__': RawTransactionsTest().main()
mit
-347,579,343,606,388,000
55.143836
263
0.641373
false
utam0k/c3os
c3os/client.py
1
1344
import socket import json import time import multiprocessing as mp from c3os import utils from c3os import conf from c3os import db from c3os.api.type import APITYPE CONF = conf.CONF def start(): """ Start client service """ mp.Process(target=client).start() def client(): """ client main routine """ db_pool = db.generate_pool() while True: send_db(db_pool) time.sleep(3.0) def send_db(db_pool): """Information on its DB is sent to other c3os. Args: db_pool (DBPool): DBPool class. Returns: None: """ all_instance_info = utils.to_dict(db_pool) all_region_names = [name for name in all_instance_info.keys()] for dest_region_name, dest in CONF['dests'].items(): host, port = dest.split(',') for region_name in all_region_names: if dest_region_name.lower() == region_name.lower(): all_instance_info.pop(region_name) break db_info = json.dumps(all_instance_info) data = json.dumps({'type': APITYPE.ADD_DB, 'body': db_info}) try: with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as sock: sock.connect((host, int(port))) sock.sendall(bytes(data, "utf-8")) except: print("Error: Connected", host, port)
mit
-5,227,906,240,231,535,000
23.888889
75
0.594494
false
plotly/python-api
packages/python/plotly/plotly/figure_factory/_annotated_heatmap.py
1
10195
from __future__ import absolute_import, division from plotly import exceptions, optional_imports import plotly.colors as clrs from plotly.figure_factory import utils from plotly.graph_objs import graph_objs from plotly.validators.heatmap import ColorscaleValidator # Optional imports, may be None for users that only use our core functionality. np = optional_imports.get_module("numpy") def validate_annotated_heatmap(z, x, y, annotation_text): """ Annotated-heatmap-specific validations Check that if a text matrix is supplied, it has the same dimensions as the z matrix. See FigureFactory.create_annotated_heatmap() for params :raises: (PlotlyError) If z and text matrices do not have the same dimensions. """ if annotation_text is not None and isinstance(annotation_text, list): utils.validate_equal_length(z, annotation_text) for lst in range(len(z)): if len(z[lst]) != len(annotation_text[lst]): raise exceptions.PlotlyError( "z and text should have the " "same dimensions" ) if x: if len(x) != len(z[0]): raise exceptions.PlotlyError( "oops, the x list that you " "provided does not match the " "width of your z matrix " ) if y: if len(y) != len(z): raise exceptions.PlotlyError( "oops, the y list that you " "provided does not match the " "length of your z matrix " ) def create_annotated_heatmap( z, x=None, y=None, annotation_text=None, colorscale="Plasma", font_colors=None, showscale=False, reversescale=False, **kwargs ): """ Function that creates annotated heatmaps This function adds annotations to each cell of the heatmap. :param (list[list]|ndarray) z: z matrix to create heatmap. :param (list) x: x axis labels. :param (list) y: y axis labels. :param (list[list]|ndarray) annotation_text: Text strings for annotations. Should have the same dimensions as the z matrix. If no text is added, the values of the z matrix are annotated. Default = z matrix values. :param (list|str) colorscale: heatmap colorscale. :param (list) font_colors: List of two color strings: [min_text_color, max_text_color] where min_text_color is applied to annotations for heatmap values < (max_value - min_value)/2. If font_colors is not defined, the colors are defined logically as black or white depending on the heatmap's colorscale. :param (bool) showscale: Display colorscale. Default = False :param (bool) reversescale: Reverse colorscale. Default = False :param kwargs: kwargs passed through plotly.graph_objs.Heatmap. These kwargs describe other attributes about the annotated Heatmap trace such as the colorscale. For more information on valid kwargs call help(plotly.graph_objs.Heatmap) Example 1: Simple annotated heatmap with default configuration >>> import plotly.figure_factory as ff >>> z = [[0.300000, 0.00000, 0.65, 0.300000], ... [1, 0.100005, 0.45, 0.4300], ... [0.300000, 0.00000, 0.65, 0.300000], ... [1, 0.100005, 0.45, 0.00000]] >>> fig = ff.create_annotated_heatmap(z) >>> fig.show() """ # Avoiding mutables in the call signature font_colors = font_colors if font_colors is not None else [] validate_annotated_heatmap(z, x, y, annotation_text) # validate colorscale colorscale_validator = ColorscaleValidator() colorscale = colorscale_validator.validate_coerce(colorscale) annotations = _AnnotatedHeatmap( z, x, y, annotation_text, colorscale, font_colors, reversescale, **kwargs ).make_annotations() if x or y: trace = dict( type="heatmap", z=z, x=x, y=y, colorscale=colorscale, showscale=showscale, reversescale=reversescale, **kwargs ) layout = dict( annotations=annotations, xaxis=dict(ticks="", dtick=1, side="top", gridcolor="rgb(0, 0, 0)"), yaxis=dict(ticks="", dtick=1, ticksuffix=" "), ) else: trace = dict( type="heatmap", z=z, colorscale=colorscale, showscale=showscale, reversescale=reversescale, **kwargs ) layout = dict( annotations=annotations, xaxis=dict( ticks="", side="top", gridcolor="rgb(0, 0, 0)", showticklabels=False ), yaxis=dict(ticks="", ticksuffix=" ", showticklabels=False), ) data = [trace] return graph_objs.Figure(data=data, layout=layout) def to_rgb_color_list(color_str, default): if "rgb" in color_str: return [int(v) for v in color_str.strip("rgb()").split(",")] elif "#" in color_str: return clrs.hex_to_rgb(color_str) else: return default def should_use_black_text(background_color): return ( background_color[0] * 0.299 + background_color[1] * 0.587 + background_color[2] * 0.114 ) > 186 class _AnnotatedHeatmap(object): """ Refer to TraceFactory.create_annotated_heatmap() for docstring """ def __init__( self, z, x, y, annotation_text, colorscale, font_colors, reversescale, **kwargs ): self.z = z if x: self.x = x else: self.x = range(len(z[0])) if y: self.y = y else: self.y = range(len(z)) if annotation_text is not None: self.annotation_text = annotation_text else: self.annotation_text = self.z self.colorscale = colorscale self.reversescale = reversescale self.font_colors = font_colors def get_text_color(self): """ Get font color for annotations. The annotated heatmap can feature two text colors: min_text_color and max_text_color. The min_text_color is applied to annotations for heatmap values < (max_value - min_value)/2. The user can define these two colors. Otherwise the colors are defined logically as black or white depending on the heatmap's colorscale. :rtype (string, string) min_text_color, max_text_color: text color for annotations for heatmap values < (max_value - min_value)/2 and text color for annotations for heatmap values >= (max_value - min_value)/2 """ # Plotly colorscales ranging from a lighter shade to a darker shade colorscales = [ "Greys", "Greens", "Blues", "YIGnBu", "YIOrRd", "RdBu", "Picnic", "Jet", "Hot", "Blackbody", "Earth", "Electric", "Viridis", "Cividis", ] # Plotly colorscales ranging from a darker shade to a lighter shade colorscales_reverse = ["Reds"] white = "#FFFFFF" black = "#000000" if self.font_colors: min_text_color = self.font_colors[0] max_text_color = self.font_colors[-1] elif self.colorscale in colorscales and self.reversescale: min_text_color = black max_text_color = white elif self.colorscale in colorscales: min_text_color = white max_text_color = black elif self.colorscale in colorscales_reverse and self.reversescale: min_text_color = white max_text_color = black elif self.colorscale in colorscales_reverse: min_text_color = black max_text_color = white elif isinstance(self.colorscale, list): min_col = to_rgb_color_list(self.colorscale[0][1], [255, 255, 255]) max_col = to_rgb_color_list(self.colorscale[-1][1], [255, 255, 255]) # swap min/max colors if reverse scale if self.reversescale: min_col, max_col = max_col, min_col if should_use_black_text(min_col): min_text_color = black else: min_text_color = white if should_use_black_text(max_col): max_text_color = black else: max_text_color = white else: min_text_color = black max_text_color = black return min_text_color, max_text_color def get_z_mid(self): """ Get the mid value of z matrix :rtype (float) z_avg: average val from z matrix """ if np and isinstance(self.z, np.ndarray): z_min = np.amin(self.z) z_max = np.amax(self.z) else: z_min = min([v for row in self.z for v in row]) z_max = max([v for row in self.z for v in row]) z_mid = (z_max + z_min) / 2 return z_mid def make_annotations(self): """ Get annotations for each cell of the heatmap with graph_objs.Annotation :rtype (list[dict]) annotations: list of annotations for each cell of the heatmap """ min_text_color, max_text_color = _AnnotatedHeatmap.get_text_color(self) z_mid = _AnnotatedHeatmap.get_z_mid(self) annotations = [] for n, row in enumerate(self.z): for m, val in enumerate(row): font_color = min_text_color if val < z_mid else max_text_color annotations.append( graph_objs.layout.Annotation( text=str(self.annotation_text[n][m]), x=self.x[m], y=self.y[n], xref="x1", yref="y1", font=dict(color=font_color), showarrow=False, ) ) return annotations
mit
-5,518,445,393,110,075,000
32.316993
87
0.567043
false
gangadhar-kadam/sapphire_app
accounts/report/budget_variance_report/budget_variance_report.py
1
4626
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. # License: GNU General Public License v3. See license.txt from __future__ import unicode_literals import webnotes from webnotes import _, msgprint from webnotes.utils import flt import time from accounts.utils import get_fiscal_year from controllers.trends import get_period_date_ranges, get_period_month_ranges def execute(filters=None): if not filters: filters = {} columns = get_columns(filters) period_month_ranges = get_period_month_ranges(filters["period"], filters["fiscal_year"]) cam_map = get_costcenter_account_month_map(filters) data = [] for cost_center, cost_center_items in cam_map.items(): for account, monthwise_data in cost_center_items.items(): row = [cost_center, account] totals = [0, 0, 0] for relevant_months in period_month_ranges: period_data = [0, 0, 0] for month in relevant_months: month_data = monthwise_data.get(month, {}) for i, fieldname in enumerate(["target", "actual", "variance"]): value = flt(month_data.get(fieldname)) period_data[i] += value totals[i] += value period_data[2] = period_data[0] - period_data[1] row += period_data totals[2] = totals[0] - totals[1] row += totals data.append(row) return columns, sorted(data, key=lambda x: (x[0], x[1])) def get_columns(filters): for fieldname in ["fiscal_year", "period", "company"]: if not filters.get(fieldname): label = (" ".join(fieldname.split("_"))).title() msgprint(_("Please specify") + ": " + label, raise_exception=True) columns = ["Cost Center:Link/Cost Center:120", "Account:Link/Account:120"] group_months = False if filters["period"] == "Monthly" else True for from_date, to_date in get_period_date_ranges(filters["period"], filters["fiscal_year"]): for label in ["Target (%s)", "Actual (%s)", "Variance (%s)"]: if group_months: label = label % (from_date.strftime("%b") + " - " + to_date.strftime("%b")) else: label = label % from_date.strftime("%b") columns.append(label+":Float:120") return columns + ["Total Target:Float:120", "Total Actual:Float:120", "Total Variance:Float:120"] #Get cost center & target details def get_costcenter_target_details(filters): return webnotes.conn.sql("""select cc.name, cc.distribution_id, cc.parent_cost_center, bd.account, bd.budget_allocated from `tabCost Center` cc, `tabBudget Detail` bd where bd.parent=cc.name and bd.fiscal_year=%s and cc.company=%s order by cc.name""" % ('%s', '%s'), (filters.get("fiscal_year"), filters.get("company")), as_dict=1) #Get target distribution details of accounts of cost center def get_target_distribution_details(filters): target_details = {} for d in webnotes.conn.sql("""select bd.name, bdd.month, bdd.percentage_allocation from `tabBudget Distribution Detail` bdd, `tabBudget Distribution` bd where bdd.parent=bd.name and bd.fiscal_year=%s""", (filters["fiscal_year"]), as_dict=1): target_details.setdefault(d.name, {}).setdefault(d.month, flt(d.percentage_allocation)) return target_details #Get actual details from gl entry def get_actual_details(filters): ac_details = webnotes.conn.sql("""select gl.account, gl.debit, gl.credit, gl.cost_center, MONTHNAME(gl.posting_date) as month_name from `tabGL Entry` gl, `tabBudget Detail` bd where gl.fiscal_year=%s and company=%s and bd.account=gl.account and bd.parent=gl.cost_center""" % ('%s', '%s'), (filters.get("fiscal_year"), filters.get("company")), as_dict=1) cc_actual_details = {} for d in ac_details: cc_actual_details.setdefault(d.cost_center, {}).setdefault(d.account, []).append(d) return cc_actual_details def get_costcenter_account_month_map(filters): import datetime costcenter_target_details = get_costcenter_target_details(filters) tdd = get_target_distribution_details(filters) actual_details = get_actual_details(filters) cam_map = {} for ccd in costcenter_target_details: for month_id in range(1, 13): month = datetime.date(2013, month_id, 1).strftime('%B') cam_map.setdefault(ccd.name, {}).setdefault(ccd.account, {})\ .setdefault(month, webnotes._dict({ "target": 0.0, "actual": 0.0 })) tav_dict = cam_map[ccd.name][ccd.account][month] month_percentage = tdd.get(ccd.distribution_id, {}).get(month, 0) \ if ccd.distribution_id else 100.0/12 tav_dict.target = flt(ccd.budget_allocated) * month_percentage / 100 for ad in actual_details.get(ccd.name, {}).get(ccd.account, []): if ad.month_name == month: tav_dict.actual += ad.debit - ad.credit return cam_map
agpl-3.0
6,781,928,573,172,004,000
35.722222
93
0.68396
false
kdeldycke/meta-package-manager
meta_package_manager/tests/test_platform.py
1
4103
# Copyright Kevin Deldycke <[email protected]> and contributors. # All Rights Reserved. # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. from types import FunctionType import pytest from ..managers import pool from ..platform import ( ALL_OS_LABELS, CURRENT_OS_ID, CURRENT_OS_LABEL, LINUX, MACOS, OS_DEFINITIONS, WINDOWS, current_os, is_linux, is_macos, is_windows, os_label, ) from .conftest import MANAGER_IDS, unless_linux, unless_macos, unless_windows def test_mutual_exclusion(): if is_linux(): assert not is_macos() assert not is_windows() assert CURRENT_OS_ID == LINUX assert CURRENT_OS_LABEL == os_label(LINUX) if is_macos(): assert not is_linux() assert not is_windows() assert CURRENT_OS_ID == MACOS assert CURRENT_OS_LABEL == os_label(MACOS) if is_windows(): assert not is_linux() assert not is_macos() assert CURRENT_OS_ID == WINDOWS assert CURRENT_OS_LABEL == os_label(WINDOWS) def test_os_definitions(): assert isinstance(OS_DEFINITIONS, dict) # Each OS definition must be unique. assert isinstance(ALL_OS_LABELS, frozenset) assert len(OS_DEFINITIONS) == len(ALL_OS_LABELS) for os_id, data in OS_DEFINITIONS.items(): # OS ID. assert isinstance(os_id, str) assert os_id assert os_id.isascii() assert os_id.isalpha() assert os_id.islower() # Metadata. assert isinstance(data, tuple) assert len(data) == 2 label, os_flag = data # OS label. assert label assert isinstance(label, str) assert label.isascii() assert label.isalpha() assert label in ALL_OS_LABELS # OS identification function. assert isinstance(os_flag, bool) os_id_func_name = f"is_{os_id}" assert os_id_func_name in globals() os_id_func = globals()[os_id_func_name] assert isinstance(os_id_func, FunctionType) assert isinstance(os_id_func(), bool) assert os_id_func() == os_flag def test_current_os_func(): # Function. os_id, label = current_os() assert os_id in OS_DEFINITIONS assert label in [os[0] for os in OS_DEFINITIONS.values()] # Constants. assert os_id == CURRENT_OS_ID assert label == CURRENT_OS_LABEL def test_os_label(): os_id, os_name = current_os() assert os_label(os_id) == os_name def test_blacklisted_manager(): """Check all managers are accounted for on each platforms.""" # TODO: Use that blacklist to generate readme.rst's support table? blacklists = { LINUX: {"cask", "mas"}, MACOS: {"apt", "flatpak", "opkg", "snap"}, WINDOWS: {"apt", "brew", "cask", "flatpak", "mas", "opkg", "snap"}, } blacklist = blacklists[current_os()[0]] # List of supported managers on the current platform. supported = {m.id for m in pool().values() if m.supported} assert supported == MANAGER_IDS - blacklist # Test unittest decorator helpers. @unless_linux def test_unless_linux(): assert is_linux() assert not is_macos() assert not is_windows() @unless_macos def test_unless_macos(): assert not is_linux() assert is_macos() assert not is_windows() @unless_windows def test_unless_windows(): assert not is_linux() assert not is_macos() assert is_windows()
gpl-2.0
920,887,995,347,523,500
28.307143
77
0.645625
false
sunqm/pyscf
pyscf/mcscf/test/test_newton_casscf.py
1
4618
#!/usr/bin/env python # Copyright 2014-2018 The PySCF Developers. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest from functools import reduce import numpy from pyscf import gto, scf, lib, fci from pyscf.mcscf import newton_casscf, CASSCF, addons mol = gto.Mole() mol.verbose = lib.logger.DEBUG mol.output = '/dev/null' mol.atom = [ ['H', ( 5.,-1. , 1. )], ['H', ( 0.,-5. ,-2. )], ['H', ( 4.,-0.5 ,-3. )], ['H', ( 0.,-4.5 ,-1. )], ['H', ( 3.,-0.5 ,-0. )], ['H', ( 0.,-3. ,-1. )], ['H', ( 2.,-2.5 , 0. )], ['H', ( 1., 1. , 3. )], ] mol.basis = 'sto-3g' mol.build() b = 1.4 mol_N2 = gto.Mole() mol_N2.build( verbose = lib.logger.DEBUG, output = '/dev/null', atom = [ ['N',( 0.000000, 0.000000, -b/2)], ['N',( 0.000000, 0.000000, b/2)], ], basis = {'N': 'ccpvdz', }, symmetry = 1 ) mf_N2 = scf.RHF (mol_N2).run () solver1 = fci.FCI(mol_N2) solver1.spin = 0 solver1.nroots = 2 solver2 = fci.FCI(mol_N2, singlet=False) solver2.spin = 2 mc_N2 = CASSCF(mf_N2, 4, 4) mc_N2 = addons.state_average_mix_(mc_N2, [solver1, solver2], (0.25,0.25,0.5)).newton () mc_N2.kernel() mf = scf.RHF(mol) mf.max_cycle = 3 mf.kernel() mc = newton_casscf.CASSCF(mf, 4, 4) mc.fcisolver = fci.direct_spin1.FCI(mol) mc.kernel() sa = CASSCF(mf, 4, 4) sa.fcisolver = fci.direct_spin1.FCI (mol) sa = sa.state_average ([0.5,0.5]).newton () sa.kernel() def tearDownModule(): global mol, mf, mc, sa, mol_N2, mf_N2, mc_N2 del mol, mf, mc, sa, mol_N2, mf_N2, mc_N2 class KnownValues(unittest.TestCase): def test_gen_g_hop(self): numpy.random.seed(1) mo = numpy.random.random(mf.mo_coeff.shape) ci0 = numpy.random.random((6,6)) ci0/= numpy.linalg.norm(ci0) gall, gop, hop, hdiag = newton_casscf.gen_g_hop(mc, mo, ci0, mc.ao2mo(mo)) self.assertAlmostEqual(lib.finger(gall), 21.288022525148595, 8) self.assertAlmostEqual(lib.finger(hdiag), -4.6864640132374618, 8) x = numpy.random.random(gall.size) u, ci1 = newton_casscf.extract_rotation(mc, x, 1, ci0) self.assertAlmostEqual(lib.finger(gop(u, ci1)), -412.9441873541524, 8) self.assertAlmostEqual(lib.finger(hop(x)), 73.358310983341198, 8) def test_get_grad(self): self.assertAlmostEqual(mc.e_tot, -3.6268060853430573, 8) self.assertAlmostEqual(abs(mc.get_grad()).max(), 0, 5) def test_sa_gen_g_hop(self): numpy.random.seed(1) mo = numpy.random.random(mf.mo_coeff.shape) ci0 = numpy.random.random((2,36)) ci0/= numpy.linalg.norm(ci0, axis=1)[:,None] ci0 = list (ci0.reshape ((2,6,6))) gall, gop, hop, hdiag = newton_casscf.gen_g_hop(sa, mo, ci0, sa.ao2mo(mo)) self.assertAlmostEqual(lib.finger(gall), 32.46973284682045, 8) self.assertAlmostEqual(lib.finger(hdiag), -63.6527761153809, 8) x = numpy.random.random(gall.size) u, ci1 = newton_casscf.extract_rotation(sa, x, 1, ci0) self.assertAlmostEqual(lib.finger(gop(u, ci1)), -49.017079186126, 8) self.assertAlmostEqual(lib.finger(hop(x)), 169.47893548740288, 8) def test_sa_get_grad(self): self.assertAlmostEqual(sa.e_tot, -3.62638372957158, 7) # MRH 06/24/2020: convergence thresh of scf may not have consistent # meaning in SA problems self.assertAlmostEqual(abs(sa.get_grad()).max(), 0, 5) def test_sa_mix(self): e = mc_N2.e_states self.assertAlmostEqual(mc_N2.e_tot, -108.80340952016508, 7) self.assertAlmostEqual(mc_N2.e_average, -108.80340952016508, 7) self.assertAlmostEqual(numpy.dot(e,[.25,.25,.5]), -108.80340952016508, 7) dm1 = mc_N2.analyze() self.assertAlmostEqual(lib.fp(dm1[0]), 0.52172669549357464, 4) self.assertAlmostEqual(lib.fp(dm1[1]), 0.53366776017869022, 4) self.assertAlmostEqual(lib.fp(dm1[0]+dm1[1]), 1.0553944556722636, 4) mc_N2.cas_natorb() if __name__ == "__main__": print("Full Tests for mcscf.addons") unittest.main()
apache-2.0
7,542,723,317,859,126,000
34.251908
82
0.621914
false
Hammer2900/SunflowerX
application/tools/find_files.py
1
13799
import os import gtk import user import pango import gobject from threading import Thread, Event class Column: ICON = 0 NAME = 1 DIRECTORY = 2 class FindFiles(gobject.GObject): """Find files tool""" __gtype_name__ = 'Sunflower_FindFiles' __gsignals__ = { 'notify-start': (gobject.SIGNAL_RUN_LAST, None, ()), 'notify-stop': (gobject.SIGNAL_RUN_LAST, None, ()) } def __init__(self, parent, application): gobject.GObject.__init__(self) # store parameters self._parent = parent self._application = application self._extensions = [] self._path = self._parent.path self._provider = None self._running = False # thread control object self._abort = Event() if hasattr(self._parent, 'get_provider'): self._provider = self._parent.get_provider() # configure window self.window = gtk.Window(type=gtk.WINDOW_TOPLEVEL) self.window.set_title(_('Find files')) self.window.set_default_size(550, 500) self.window.set_position(gtk.WIN_POS_CENTER_ON_PARENT) self.window.set_transient_for(application) self.window.set_border_width(7) self.window.set_type_hint(gtk.gdk.WINDOW_TYPE_HINT_DIALOG) self.window.set_wmclass('Sunflower', 'Sunflower') self.window.connect('key-press-event', self._handle_key_press) # create interface vbox = gtk.VBox(False, 7) # create path and basic options self._table_basic = gtk.Table(3, 2, False) self._table_basic.set_col_spacings(5) self._table_basic.set_row_spacings(2) label_path = gtk.Label(_('Search in:')) label_path.set_alignment(0, 0.5) self._entry_path = gtk.Entry() self._entry_path.connect('activate', self.find_files) if hasattr(self._parent, 'path'): # get path from the parent self._entry_path.set_text(self._parent.path) else: # parent has no path, set user home directory self._entry_path.set_text(os.path.expanduser(user.home)) button_browse = gtk.Button(label=_('Browse')) button_browse.connect('clicked', self._choose_directory) self._checkbox_recursive = gtk.CheckButton(label=_('Search recursively')) self._checkbox_recursive.set_active(True) # create extensions notebook self._extension_list = gtk.Notebook() # create list self._list = gtk.ListStore(str, str, str) self._names = gtk.TreeView(model=self._list) cell_icon = gtk.CellRendererPixbuf() cell_name = gtk.CellRendererText() cell_directory = gtk.CellRendererText() col_name = gtk.TreeViewColumn(_('Name')) col_name.set_expand(True) col_directory = gtk.TreeViewColumn(_('Location')) col_directory.set_expand(True) # pack renderer col_name.pack_start(cell_icon, False) col_name.pack_start(cell_name, True) col_directory.pack_start(cell_directory, True) # connect renderer attributes col_name.add_attribute(cell_icon, 'icon-name', Column.ICON) col_name.add_attribute(cell_name, 'text', Column.NAME) col_directory.add_attribute(cell_directory, 'text', Column.DIRECTORY) self._names.append_column(col_name) self._names.append_column(col_directory) self._names.connect('row-activated', self.__handle_row_activated) container = gtk.ScrolledWindow() container.set_policy(gtk.POLICY_AUTOMATIC, gtk.POLICY_ALWAYS) container.set_shadow_type(gtk.SHADOW_IN) # create status label self._status = gtk.Label() self._status.set_alignment(0, 0.5) self._status.set_ellipsize(pango.ELLIPSIZE_MIDDLE) self._status.set_property('no-show-all', True) # create controls hbox_controls = gtk.HBox(False, 5) self._image_find = gtk.Image() self._image_find.set_from_stock(gtk.STOCK_MEDIA_PLAY, gtk.ICON_SIZE_BUTTON) self._button_find = gtk.Button() self._button_find.set_label(_('Start')) self._button_find.set_image(self._image_find) self._button_find.connect('clicked', self.find_files) button_close = gtk.Button(stock=gtk.STOCK_CLOSE) button_close.connect('clicked', self._close_window) # pack interface self._table_basic.attach(label_path, 0, 1, 0, 1, xoptions=gtk.SHRINK | gtk.FILL) self._table_basic.attach(self._entry_path, 1, 2, 0, 1, xoptions=gtk.EXPAND | gtk.FILL) self._table_basic.attach(button_browse, 2, 3, 0, 1, xoptions=gtk.SHRINK | gtk.FILL) self._table_basic.attach(self._checkbox_recursive, 1, 2, 1, 2) container.add(self._names) hbox_controls.pack_end(self._button_find, False, False, 0) hbox_controls.pack_end(button_close, False, False, 0) vbox.pack_start(self._table_basic, False, False, 0) vbox.pack_start(self._extension_list, False, False, 0) vbox.pack_end(hbox_controls, False, False, 0) vbox.pack_end(self._status, False, False, 0) vbox.pack_end(container, True, True, 0) self.window.add(vbox) # create extensions self.__create_extensions() # show all widgets self.window.show_all() def __handle_row_activated(self, treeview, path, view_column, data=None): """Handle actions on list""" # get list selection selection = treeview.get_selection() list_, iter_ = selection.get_selected() # we need selection for this if iter_ is None: return name = list_.get_value(iter_, Column.NAME) path = list_.get_value(iter_, Column.DIRECTORY) # get active object active_object = self._application.get_active_object() if hasattr(active_object, 'change_path'): # change path active_object.change_path(path, name) # close window self._close_window() else: # notify user about active object dialog = gtk.MessageDialog( self.window, gtk.DIALOG_DESTROY_WITH_PARENT, gtk.MESSAGE_INFO, gtk.BUTTONS_OK, _( 'Active object doesn\'t support changing ' 'path. Set focus on a different object, ' 'preferably file list, and try again.' ) ) dialog.run() dialog.destroy() def __create_extensions(self): """Create rename extensions""" for ExtensionClass in self._application.find_extension_classes.values(): extension = ExtensionClass(self) title = extension.get_title() # add tab self._extension_list.append_page(extension.get_container(), gtk.Label(title)) # store extension for later use self._extensions.append(extension) def __update_status_label(self, path): """Update status label with current scanning path""" self._status.set_text(path) def __update_status(self, running=True): """Update button status""" self._running = running if running: # disable interface to prevent changes during search self._table_basic.set_sensitive(False) self._extension_list.set_sensitive(False) # show status bar self._status.show() # update find button self._image_find.set_from_stock(gtk.STOCK_MEDIA_STOP, gtk.ICON_SIZE_BUTTON) self._button_find.set_label(_('Stop')) else: # enable interface to prevent changes during search self._table_basic.set_sensitive(True) self._extension_list.set_sensitive(True) # hide status bar self._status.hide() # update find button self._image_find.set_from_stock(gtk.STOCK_MEDIA_PLAY, gtk.ICON_SIZE_BUTTON) self._button_find.set_label(_('Start')) def __find_files(self, path, children, scan_recursively): """Threaded find files method""" scan_queue = [] extension_list = [] # prepare extension objects for operation for child in children: extension_list.append(child.get_data('extension')) # tell extensions search is starting self.emit('notify-start') # update thread status gobject.idle_add(self.__update_status, True) gobject.idle_add(self.__update_status_label, path) # add current path to scan queue try: item_list = self._provider.list_dir(path) item_list = map(lambda new_item: os.path.join(path, new_item), item_list) scan_queue.extend(item_list) except: pass # traverse through directories while not self._abort.is_set() and len(scan_queue) > 0: # get next item in queue item = scan_queue.pop(0) if self._provider.is_dir(item) and scan_recursively: # extend scan queue with directory content gobject.idle_add(self.__update_status_label, item) try: item_list = self._provider.list_dir(item) item_list = map(lambda new_item: os.path.join(item, new_item), item_list) scan_queue.extend(item_list) except: pass # check if item fits cirteria match = True for extension in extension_list: if not extension.is_path_ok(item): match = False break # add item if score is right if match: name = os.path.basename(item) path = os.path.dirname(item) icon = self._application.icon_manager.get_icon_for_file(item) self._list.append((icon, name, path)) # update thread status gobject.idle_add(self.__update_status, False) # tell extensions search has been stopped self.emit('notify-stop') def _close_window(self, widget=None, data=None): """Close window""" self._abort.set() # notify search thread we are terminating self.window.destroy() def _choose_directory(self, widget=None, data=None): """Show 'FileChooser' dialog""" dialog = gtk.FileChooserDialog( title=_('Find files'), parent=self._application, action=gtk.FILE_CHOOSER_ACTION_SELECT_FOLDER, buttons=( gtk.STOCK_CANCEL, gtk.RESPONSE_REJECT, gtk.STOCK_OK, gtk.RESPONSE_ACCEPT ) ) dialog.set_filename(self._entry_path.get_text()) response = dialog.run() if response == gtk.RESPONSE_ACCEPT: self._entry_path.set_text(dialog.get_filename()) dialog.destroy() def _handle_key_press(self, widget, event, data=None): """Handle pressing keys""" if event.keyval == gtk.keysyms.Escape: self._close_window() def stop_search(self, widget=None, data=None): """Stop searching for files""" pass def find_files(self, widget=None, data=None): """Start searching for files""" if not self._running: # thread is not running, start it path = self._entry_path.get_text() # make sure we have a valid provider if self._provider is None: ProviderClass = self._application.get_provider_by_protocol('file') self._provider = ProviderClass(self._parent) # check if specified path exists if not self._provider.is_dir(path): dialog = gtk.MessageDialog( self.window, gtk.DIALOG_DESTROY_WITH_PARENT, gtk.MESSAGE_ERROR, gtk.BUTTONS_OK, _( 'Specified path is not valid or doesn\'t ' 'exist anymore. Please check your selection ' 'and try again.' ) ) dialog.run() dialog.destroy() return # get list of active extensions active_children = filter( lambda child: child.get_data('extension').is_active(), self._extension_list.get_children() ) if len(active_children) == 0: dialog = gtk.MessageDialog( self.window, gtk.DIALOG_DESTROY_WITH_PARENT, gtk.MESSAGE_WARNING, gtk.BUTTONS_OK, _( 'You need to enable at least one extension ' 'in order to find files and directories!' ) ) dialog.run() dialog.destroy() return # set thread control objects self._abort.clear() # clear existing list self._list.clear() # start the thread params = { 'path': path, 'children': active_children, 'scan_recursively': self._checkbox_recursive.get_active() } thread = Thread(target=self.__find_files, kwargs=params) thread.start() else: # thread is running, set abort event self._abort.set()
gpl-3.0
-399,886,151,447,045,440
32.250602
94
0.560838
false
bdang2012/taiga-back-casting
tests/unit/test_timeline.py
1
3510
# Copyright (C) 2014-2015 Andrey Antukh <[email protected]> # Copyright (C) 2014-2015 Jesús Espino <[email protected]> # Copyright (C) 2014-2015 David Barragán <[email protected]> # Copyright (C) 2014-2015 Anler Hernández <[email protected]> # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from unittest.mock import patch, call from django.core.exceptions import ValidationError from taiga.timeline import service from taiga.timeline.models import Timeline from taiga.projects.models import Project from taiga.users.models import User import pytest def test_push_to_timeline_many_objects(): with patch("taiga.timeline.service._add_to_object_timeline") as mock: users = [User(), User(), User()] project = Project() service.push_to_timeline(users, project, "test", project.created_date) assert mock.call_count == 3 assert mock.mock_calls == [ call(users[0], project, "test", project.created_date, "default", {}), call(users[1], project, "test", project.created_date, "default", {}), call(users[2], project, "test", project.created_date, "default", {}), ] with pytest.raises(Exception): service.push_to_timeline(None, project, "test") def test_add_to_objects_timeline(): with patch("taiga.timeline.service._add_to_object_timeline") as mock: users = [User(), User(), User()] project = Project() service._add_to_objects_timeline(users, project, "test", project.created_date) assert mock.call_count == 3 assert mock.mock_calls == [ call(users[0], project, "test", project.created_date, "default", {}), call(users[1], project, "test", project.created_date, "default", {}), call(users[2], project, "test", project.created_date, "default", {}), ] with pytest.raises(Exception): service.push_to_timeline(None, project, "test") def test_get_impl_key_from_model(): assert service._get_impl_key_from_model(Timeline, "test") == "timeline.timeline.test" with pytest.raises(Exception): service._get_impl_key(None) def test_get_impl_key_from_typename(): assert service._get_impl_key_from_typename("timeline.timeline", "test") == "timeline.timeline.test" with pytest.raises(Exception): service._get_impl_key(None) def test_register_timeline_implementation(): test_func = lambda x: "test-func-result" service.register_timeline_implementation("timeline.timeline", "test", test_func) assert service._timeline_impl_map["timeline.timeline.test"](None) == "test-func-result" @service.register_timeline_implementation("timeline.timeline", "test-decorator") def decorated_test_function(x): return "test-decorated-func-result" assert service._timeline_impl_map["timeline.timeline.test-decorator"](None) == "test-decorated-func-result"
agpl-3.0
-5,399,271,881,389,798,000
42.296296
111
0.687482
false
trnewman/VT-USRP-daughterboard-drivers_python
gr-wxgui/src/python/numbersink.py
1
25273
#!/usr/bin/env python # # Copyright 2003,2004,2005,2006,2007 Free Software Foundation, Inc. # # This file is part of GNU Radio # # GNU Radio is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3, or (at your option) # any later version. # # GNU Radio is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with GNU Radio; see the file COPYING. If not, write to # the Free Software Foundation, Inc., 59 Temple Place - Suite 330, # Boston, MA 02111-1307, USA. # from gnuradio import gr, gru, window from gnuradio.wxgui import stdgui import wx #from wx import StaticText import gnuradio.wxgui.plot as plot import numpy import threading import math default_numbersink_size = (640,240) default_number_rate = gr.prefs().get_long('wxgui', 'number_rate', 15) class number_sink_base(object): def __init__(self, input_is_real=False, unit='',base_value=0, minval=-100.0,maxval=100.0,factor=1.0,decimal_places=10, ref_level=50, sample_rate=1, number_rate=default_number_rate, average=False, avg_alpha=None, label='', peak_hold=False): # initialize common attributes self.unit=unit self.base_value = base_value self.minval=minval self.maxval=maxval self.factor=factor self.y_divs = 8 self.decimal_places=decimal_places self.ref_level = ref_level self.sample_rate = sample_rate number_size=1 self.number_size = number_size self.number_rate = number_rate self.average = average if avg_alpha is None: self.avg_alpha = 2.0 / number_rate else: self.avg_alpha = avg_alpha self.label = label self.peak_hold = peak_hold self.show_gauge = True self.input_is_real = input_is_real self.msgq = gr.msg_queue(2) # queue that holds a maximum of 2 messages def set_decimal_places(self, decimal_places): self.decimal_places = decimal_places def set_ref_level(self, ref_level): self.ref_level = ref_level def print_current_value(self, comment): print comment,self.win.current_value def set_average(self, average): self.average = average if average: self.avg.set_taps(self.avg_alpha) self.set_peak_hold(False) else: self.avg.set_taps(1.0) def set_peak_hold(self, enable): self.peak_hold = enable if enable: self.set_average(False) self.win.set_peak_hold(enable) def set_show_gauge(self, enable): self.show_gauge = enable self.win.set_show_gauge(enable) def set_avg_alpha(self, avg_alpha): self.avg_alpha = avg_alpha def set_base_value(self, base_value): self.base_value = base_value class number_sink_f(gr.hier_block, number_sink_base): def __init__(self, fg, parent, unit='',base_value=0,minval=-100.0,maxval=100.0,factor=1.0, decimal_places=10, ref_level=50, sample_rate=1, #number_size=512, number_rate=default_number_rate, average=False, avg_alpha=None, label='', size=default_numbersink_size, peak_hold=False): number_sink_base.__init__(self, unit=unit, input_is_real=True, base_value=base_value, minval=minval,maxval=maxval,factor=factor, decimal_places=decimal_places, ref_level=ref_level, sample_rate=sample_rate, #number_size=number_size, number_rate=number_rate, average=average, avg_alpha=avg_alpha, label=label, peak_hold=peak_hold) number_size=1 #s2p = gr.stream_to_vector(gr.sizeof_float, number_size) one_in_n = gr.keep_one_in_n(gr.sizeof_float, max(1, int(sample_rate/number_rate))) #c2mag = gr.complex_to_mag(number_size) self.avg = gr.single_pole_iir_filter_ff(1.0, number_size) # FIXME We need to add 3dB to all bins but the DC bin #log = gr.nlog10_ff(20, number_size, # -20*math.log10(number_size)-10*math.log10(power/number_size)) sink = gr.message_sink(gr.sizeof_float , self.msgq, True) #fg.connect (s2p, one_in_n, fft, c2mag, self.avg, log, sink) fg.connect(self.avg,one_in_n,sink) gr.hier_block.__init__(self, fg, self.avg, sink) self.win = number_window(self, parent, size=size,label=label) self.set_average(self.average) class number_sink_c(gr.hier_block, number_sink_base): def __init__(self, fg, parent, unit='',base_value=0,minval=-100.0,maxval=100.0,factor=1.0, decimal_places=10, ref_level=50, sample_rate=1, #number_size=512, number_rate=default_number_rate, average=False, avg_alpha=None, label='', size=default_numbersink_size, peak_hold=False): number_sink_base.__init__(self, unit=unit, input_is_real=False, base_value=base_value,factor=factor, minval=minval,maxval=maxval,decimal_places=decimal_places, ref_level=ref_level, sample_rate=sample_rate, #number_size=number_size, number_rate=number_rate, average=average, avg_alpha=avg_alpha, label=label, peak_hold=peak_hold) number_size=1 one_in_n = gr.keep_one_in_n(gr.sizeof_gr_complex, max(1, int(sample_rate/number_rate))) #c2mag = gr.complex_to_mag(number_size) self.avg = gr.single_pole_iir_filter_cc(1.0, number_size) # FIXME We need to add 3dB to all bins but the DC bin #log = gr.nlog10_ff(20, number_size, # -20*math.log10(number_size)-10*math.log10(power/number_size)) sink = gr.message_sink(gr.sizeof_gr_complex , self.msgq, True) #fg.connect (s2p, one_in_n, fft, c2mag, self.avg, log, sink) fg.connect(self.avg,one_in_n,sink) gr.hier_block.__init__(self, fg, self.avg, sink) self.win = number_window(self, parent, size=size,label=label) self.set_average(self.average) # ------------------------------------------------------------------------ myDATA_EVENT = wx.NewEventType() EVT_DATA_EVENT = wx.PyEventBinder (myDATA_EVENT, 0) class DataEvent(wx.PyEvent): def __init__(self, data): wx.PyEvent.__init__(self) self.SetEventType (myDATA_EVENT) self.data = data def Clone (self): self.__class__ (self.GetId()) class input_watcher (threading.Thread): def __init__ (self, msgq, number_size, event_receiver, **kwds): threading.Thread.__init__ (self, **kwds) self.setDaemon (1) self.msgq = msgq self.number_size = number_size self.event_receiver = event_receiver self.keep_running = True self.start () def run (self): while (self.keep_running): msg = self.msgq.delete_head() # blocking read of message queue itemsize = int(msg.arg1()) nitems = int(msg.arg2()) s = msg.to_string() # get the body of the msg as a string # There may be more than one number in the message. # If so, we take only the last one if nitems > 1: start = itemsize * (nitems - 1) s = s[start:start+itemsize] complex_data = numpy.fromstring (s, numpy.float32) de = DataEvent (complex_data) wx.PostEvent (self.event_receiver, de) del de #======================================================================================== class static_text_window (wx.StaticText): #plot.PlotCanvas): def __init__ (self, parent, numbersink,id = -1,label="number", pos = wx.DefaultPosition, size = wx.DefaultSize, style = wx.DEFAULT_FRAME_STYLE, name = ""): #plot.PlotCanvas.__init__ (self, parent, id, pos, size, style, name) wx.StaticText.__init__(self, parent, id, label, pos, size, style, name) #self.static_text=wx.StaticText( parent, id, label, pos, (size[0]/2,size[1]/2), style, name) #gauge_style = wx.GA_HORIZONTAL #self.gauge=wx.Gauge( parent, id, range=1000, pos=(pos[0],pos[1]+size[1]/2),size=(size[0]/2,size[1]/2), style=gauge_style, name = "gauge") #wx.BoxSizer.__init__ (self,wx.VERTICAL) #self.Add (self.static_text, 0, wx.EXPAND) #self.Add (self.gauge, 1, wx.EXPAND) self.parent=parent self.label=label #self.y_range = None self.numbersink = numbersink self.peak_hold = False self.peak_vals = None #self.SetEnableGrid (True) # self.SetEnableZoom (True) # self.SetBackgroundColour ('black') self.build_popup_menu() #EVT_DATA_EVENT (self, self.set_data) #wx.EVT_CLOSE (self, self.on_close_window) #self.Bind(wx.EVT_RIGHT_UP, self.on_right_click) self.Bind(wx.EVT_RIGHT_UP, self.on_right_click) #self.input_watcher = input_watcher(numbersink.msgq, numbersink.number_size, self) def on_close_window (self, event): print "number_window:on_close_window" self.keep_running = False def set_peak_hold(self, enable): self.peak_hold = enable self.peak_vals = None def update_y_range (self): ymax = self.numbersink.ref_level ymin = self.numbersink.ref_level - self.numbersink.decimal_places * self.numbersink.y_divs self.y_range = self._axisInterval ('min', ymin, ymax) def on_average(self, evt): # print "on_average" self.numbersink.set_average(evt.IsChecked()) def on_peak_hold(self, evt): # print "on_peak_hold" self.numbersink.set_peak_hold(evt.IsChecked()) def on_show_gauge(self, evt): # print "on_show_gauge" #if evt.IsChecked(): self.numbersink.set_show_gauge(evt.IsChecked()) print evt.IsChecked() # print "show gauge" #else: # self.parent.gauge.Hide() # print "hide gauge" def on_incr_ref_level(self, evt): # print "on_incr_ref_level" self.numbersink.set_ref_level(self.numbersink.ref_level + self.numbersink.decimal_places) def on_decr_ref_level(self, evt): # print "on_decr_ref_level" self.numbersink.set_ref_level(self.numbersink.ref_level - self.numbersink.decimal_places) def on_incr_decimal_places(self, evt): # print "on_incr_decimal_places" self.numbersink.set_decimal_places(self.numbersink.decimal_places+1) #next_up(self.numbersink.decimal_places, (1,2,5,10,20))) def on_decr_decimal_places(self, evt): # print "on_decr_decimal_places" self.numbersink.set_decimal_places(max(self.numbersink.decimal_places-1,0)) #next_down(self.numbersink.decimal_places, (1,2,5,10,20))) def on_decimal_places(self, evt): # print "on_decimal_places" Id = evt.GetId() if Id == self.id_decimal_places_0: self.numbersink.set_decimal_places(0) elif Id == self.id_decimal_places_1: self.numbersink.set_decimal_places(1) elif Id == self.id_decimal_places_2: self.numbersink.set_decimal_places(2) elif Id == self.id_decimal_places_3: self.numbersink.set_decimal_places(3) elif Id == self.id_decimal_places_6: self.numbersink.set_decimal_places(6) elif Id == self.id_decimal_places_9: self.numbersink.set_decimal_places(9) def on_right_click(self, event): menu = self.popup_menu for id, pred in self.checkmarks.items(): item = menu.FindItemById(id) item.Check(pred()) self.PopupMenu(menu, event.GetPosition()) def build_popup_menu(self): #self.id_hide_gauge = wx.NewId() self.id_show_gauge = wx.NewId() self.id_incr_ref_level = wx.NewId() self.id_decr_ref_level = wx.NewId() self.id_incr_decimal_places = wx.NewId() self.id_decr_decimal_places = wx.NewId() self.id_decimal_places_0 = wx.NewId() self.id_decimal_places_1 = wx.NewId() self.id_decimal_places_2 = wx.NewId() self.id_decimal_places_3 = wx.NewId() self.id_decimal_places_6 = wx.NewId() self.id_decimal_places_9 = wx.NewId() self.id_average = wx.NewId() self.id_peak_hold = wx.NewId() self.Bind(wx.EVT_MENU, self.on_average, id=self.id_average) self.Bind(wx.EVT_MENU, self.on_peak_hold, id=self.id_peak_hold) #self.Bind(wx.EVT_MENU, self.on_hide_gauge, id=self.id_hide_gauge) self.Bind(wx.EVT_MENU, self.on_show_gauge, id=self.id_show_gauge) self.Bind(wx.EVT_MENU, self.on_incr_ref_level, id=self.id_incr_ref_level) self.Bind(wx.EVT_MENU, self.on_decr_ref_level, id=self.id_decr_ref_level) self.Bind(wx.EVT_MENU, self.on_incr_decimal_places, id=self.id_incr_decimal_places) self.Bind(wx.EVT_MENU, self.on_decr_decimal_places, id=self.id_decr_decimal_places) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_0) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_1) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_2) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_3) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_6) self.Bind(wx.EVT_MENU, self.on_decimal_places, id=self.id_decimal_places_9) # make a menu menu = wx.Menu() self.popup_menu = menu menu.AppendCheckItem(self.id_average, "Average") menu.AppendCheckItem(self.id_peak_hold, "Peak Hold") #menu.Append(self.id_hide_gauge, "Hide gauge") menu.AppendCheckItem(self.id_show_gauge, "Show gauge") menu.Append(self.id_incr_ref_level, "Incr Ref Level") menu.Append(self.id_decr_ref_level, "Decr Ref Level") menu.Append(self.id_incr_decimal_places, "Incr decimal places") menu.Append(self.id_decr_decimal_places, "Decr decimal places") menu.AppendSeparator() # we'd use RadioItems for these, but they're not supported on Mac menu.AppendCheckItem(self.id_decimal_places_0, "0 decimal places") menu.AppendCheckItem(self.id_decimal_places_1, "1 decimal places") menu.AppendCheckItem(self.id_decimal_places_2, "2 decimal places") menu.AppendCheckItem(self.id_decimal_places_3, "3 decimal places") menu.AppendCheckItem(self.id_decimal_places_6, "6 decimal places") menu.AppendCheckItem(self.id_decimal_places_9, "9 decimal places") self.checkmarks = { self.id_average : lambda : self.numbersink.average, self.id_peak_hold : lambda : self.numbersink.peak_hold,# self.id_hide_gauge : lambda : self.numbersink.hide_gauge, self.id_show_gauge : lambda : self.numbersink.show_gauge, self.id_decimal_places_0 : lambda : self.numbersink.decimal_places == 0, self.id_decimal_places_1 : lambda : self.numbersink.decimal_places == 1, self.id_decimal_places_2 : lambda : self.numbersink.decimal_places == 2, self.id_decimal_places_3 : lambda : self.numbersink.decimal_places == 3, self.id_decimal_places_6 : lambda : self.numbersink.decimal_places == 6, self.id_decimal_places_9 : lambda : self.numbersink.decimal_places == 9, } def next_up(v, seq): """ Return the first item in seq that is > v. """ for s in seq: if s > v: return s return v def next_down(v, seq): """ Return the last item in seq that is < v. """ rseq = list(seq[:]) rseq.reverse() for s in rseq: if s < v: return s return v #======================================================================================== class number_window (plot.PlotCanvas): def __init__ (self, numbersink, parent, id = -1,label="number", pos = wx.DefaultPosition, size = wx.DefaultSize, style = wx.DEFAULT_FRAME_STYLE, name = ""): plot.PlotCanvas.__init__ (self, parent, id, pos, size, style, name) #wx.StaticText.__init__(self, parent, id, label, pos, (size[0]/2,size[1]/2), style, name) #print 'parent',parent self.static_text=static_text_window( self, numbersink,id, label, pos, (size[0]/2,size[1]/2), style, name) gauge_style = wx.GA_HORIZONTAL vbox=wx.BoxSizer(wx.VERTICAL) vbox.Add (self.static_text, 0, wx.EXPAND) self.current_value=None if numbersink.input_is_real: self.gauge=wx.Gauge( self, id, range=1000, pos=(pos[0],pos[1]+size[1]/2),size=(size[0]/2,size[1]/2), style=gauge_style, name = "gauge") vbox.Add (self.gauge, 1, wx.EXPAND) else: self.gauge=wx.Gauge( self, id, range=1000, pos=(pos[0],pos[1]+size[1]/3),size=(size[0]/2,size[1]/3), style=gauge_style, name = "gauge") #hbox=wx.BoxSizer(wx.HORIZONTAL) self.gauge_imag=wx.Gauge( self, id, range=1000, pos=(pos[0],pos[1]+size[1]*2/3),size=(size[0]/2,size[1]/3), style=gauge_style, name = "gauge_imag") vbox.Add (self.gauge, 1, wx.EXPAND) vbox.Add (self.gauge_imag, 1, wx.EXPAND) #vbox.Add (hbox, 1, wx.EXPAND) self.sizer = vbox self.SetSizer (self.sizer) self.SetAutoLayout (True) self.sizer.Fit (self) self.label=label #self.y_range = None self.numbersink = numbersink self.peak_hold = False self.peak_vals = None #self.SetEnableGrid (True) # self.SetEnableZoom (True) # self.SetBackgroundColour ('black') #self.build_popup_menu() EVT_DATA_EVENT (self, self.set_data) wx.EVT_CLOSE (self, self.on_close_window) #self.Bind(wx.EVT_RIGHT_UP, self.on_right_click) #self.Bind(wx.EVT_RIGHT_UP, self.on_right_click) self.input_watcher = input_watcher(numbersink.msgq, numbersink.number_size, self) def on_close_window (self, event): print "number_window:on_close_window" self.keep_running = False def set_show_gauge(self, enable): self.show_gauge = enable if enable: self.gauge.Show() if not self.numbersink.input_is_real: self.gauge_imag.Show() #print 'show' else: self.gauge.Hide() if not self.numbersink.input_is_real: self.gauge_imag.Hide() #print 'hide' def set_data (self, evt): numbers = evt.data L = len (numbers) if self.peak_hold: if self.peak_vals is None: self.peak_vals = numbers else: self.peak_vals = numpy.maximum(numbers, self.peak_vals) numbers = self.peak_vals if self.numbersink.input_is_real: real_value=numbers[0]*self.numbersink.factor + self.numbersink.base_value imag_value=0.0 self.current_value=real_value else: real_value=numbers[0]*self.numbersink.factor + self.numbersink.base_value imag_value=numbers[1]*self.numbersink.factor + self.numbersink.base_value self.current_value=complex(real_value,imag_value) #x = max(abs(self.numbersink.sample_rate), abs(self.numbersink.base_value)) x = max(real_value, imag_value) if x >= 1e9: sf = 1e-9 unit_prefix = "G" elif x >= 1e6: sf = 1e-6 unit_prefix = "M" elif x>= 1e3: sf = 1e-3 unit_prefix = "k" else : sf = 1 unit_prefix = "" #self.update_y_range () if self.numbersink.input_is_real: showtext = "%s: %.*f %s%s" % (self.label, self.numbersink.decimal_places,real_value*sf,unit_prefix,self.numbersink.unit) else: showtext = "%s: %.*f,%.*f %s%s" % (self.label, self.numbersink.decimal_places,real_value*sf, self.numbersink.decimal_places,imag_value*sf,unit_prefix,self.numbersink.unit) self.static_text.SetLabel(showtext) #print (int(float((real_value-self.numbersink.base_value)*1000.0/(self.numbersink.maxval-self.numbersink.minval)))+500) self.gauge.SetValue(int(float((real_value-self.numbersink.base_value)*1000.0/(self.numbersink.maxval-self.numbersink.minval)))+500) if not self.numbersink.input_is_real: self.gauge.SetValue(int(float((imag_value-self.numbersink.base_value)*1000.0/(self.numbersink.maxval-self.numbersink.minval)))+500) def set_peak_hold(self, enable): self.peak_hold = enable self.peak_vals = None def update_y_range (self): ymax = self.numbersink.ref_level ymin = self.numbersink.ref_level - self.numbersink.decimal_places * self.numbersink.y_divs self.y_range = self._axisInterval ('min', ymin, ymax) def on_average(self, evt): # print "on_average" self.numbersink.set_average(evt.IsChecked()) def on_peak_hold(self, evt): # print "on_peak_hold" self.numbersink.set_peak_hold(evt.IsChecked()) # ---------------------------------------------------------------- # Deprecated interfaces # ---------------------------------------------------------------- # returns (block, win). # block requires a single input stream of float # win is a subclass of wxWindow def make_number_sink_f(fg, parent, label, number_size, input_rate, ymin = 0, ymax=50): block = number_sink_f(fg, parent, label=label, number_size=number_size, sample_rate=input_rate, decimal_places=(ymax - ymin)/8, ref_level=ymax) return (block, block.win) # returns (block, win). # block requires a single input stream of gr_complex # win is a subclass of wxWindow def make_number_sink_c(fg, parent, label, number_size, input_rate, ymin=0, ymax=50): block = number_sink_c(fg, parent, label=label, number_size=number_size, sample_rate=input_rate, decimal_places=(ymax - ymin)/8, ref_level=ymax) return (block, block.win) # ---------------------------------------------------------------- # Standalone test app # ---------------------------------------------------------------- class test_app_flow_graph (stdgui.gui_flow_graph): def __init__(self, frame, panel, vbox, argv): stdgui.gui_flow_graph.__init__ (self, frame, panel, vbox, argv) #number_size = 256 # build our flow graph input_rate = 20.48e3 # Generate a complex sinusoid src1 = gr.sig_source_c (input_rate, gr.GR_SIN_WAVE, 2e3, 1) #src1 = gr.sig_source_c (input_rate, gr.GR_CONST_WAVE, 5.75e3, 1) # We add these throttle blocks so that this demo doesn't # suck down all the CPU available. Normally you wouldn't use these. thr1 = gr.throttle(gr.sizeof_gr_complex, input_rate) #sink1 = number_sink_c (self, panel, label="Complex Data", number_size=number_size, # sample_rate=input_rate, base_value=100e3, # ref_level=0, decimal_places=3) #vbox.Add (sink1.win, 1, wx.EXPAND) #self.connect (src1, thr1, sink1) src2 = gr.sig_source_f (input_rate, gr.GR_SIN_WAVE, 2e3, 1) #src2 = gr.sig_source_f (input_rate, gr.GR_CONST_WAVE, 5.75e3, 1) thr2 = gr.throttle(gr.sizeof_float, input_rate) sink2 = number_sink_f (self, panel, unit='Hz',label="Real Data", avg_alpha=0.001,#number_size=number_size*2, sample_rate=input_rate, base_value=100e3, ref_level=0, decimal_places=3) vbox.Add (sink2.win, 1, wx.EXPAND) sink3 = number_sink_c (self, panel, unit='V',label="Complex Data", avg_alpha=0.001,#number_size=number_size*2, sample_rate=input_rate, base_value=0, ref_level=0, decimal_places=3) vbox.Add (sink3.win, 1, wx.EXPAND) self.connect (src2, thr2, sink2) self.connect (src1, thr1, sink3) def main (): app = stdgui.stdapp (test_app_flow_graph, "Number Sink Test App") app.MainLoop () if __name__ == '__main__': main ()
gpl-3.0
1,420,374,950,807,376,000
40.161238
158
0.585328
false
icebreaker/pyGLox
demo/particlesystem.py
1
1664
""" Copyright (c) 2011, Mihail Szabolcs All rights reserved. See LICENSE for more information. """ import random import math from pyglet.gl import * class Particle(object): def __init__(self): self.p = [0,0,0] self.a = 1 self.dx = (random.random() - 0.5) self.dy = (random.random() - 0.5) def update(self, dt): self.p[0] += self.dx * dt self.p[1] += math.fabs(self.dy / 3 * dt) self.a -= math.fabs(self.dx * 4) * dt self.a -= math.fabs(self.dy / 2) * dt if self.a <= 0: self.p = [0,0,0] self.a = 1 self.dx = (random.random() - 0.5) self.dy = (random.random() - 0.5) def draw(self): #glColor4f(1, 0.6, 0.0, self.a) glColor4f(0.65, 0.0, 0.15, self.a) glVertex3f(self.p[0], self.p[1], self.p[2]) class ParticleSystem(object): def __init__(self, texture, n=512, p=Particle): self.texture = texture self.n = n self.particles = [] for i in range(n): self.particles.append(p()) def update(self, dt): for i in range(self.n): self.particles[i].update(dt) def draw(self): self.texture.bind() glEnable(GL_BLEND) glBlendFunc(GL_SRC_ALPHA, GL_ONE) glEnable(GL_POINT_SPRITE) glPointSize(60.0) glTexEnvi(GL_POINT_SPRITE, GL_COORD_REPLACE, GL_TRUE); #glPointParameterf(GL_POINT_FADE_THRESHOLD_SIZE, 60.0); #glPointParameterf(GL_POINT_SIZE_MIN, 30.0); #glPointParameterf(GL_POINT_SIZE_MAX, 40.0); glDisable(GL_DEPTH_TEST) glDepthMask(GL_FALSE) for i in range(self.n): glBegin(GL_POINTS) self.particles[i].draw() glEnd() glDepthMask(GL_TRUE) glEnable(GL_DEPTH_TEST) glDisable(GL_POINT_SPRITE) glDisable(GL_BLEND) self.texture.release() glColor4f(1,1,1,1)
mit
7,980,364,552,582,119,000
21.794521
57
0.640625
false
dnaextrim/django_adminlte_x
adminlte/static/plugins/datatables/extensions/Responsive/examples/initialisation/index.html.py
1
2778
XXXXXXXXX XXXXX XXXXXX XXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXXXXXXX X XXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXX XXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXXXXX XXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXX XXXXXXXXXXXXX XXX XX XXX XX X XXXXXXXXX XX X XXXXXX XX XXXXXXXXX XXXXXXXXX XXX XXXXXXXXXXXXXXXXX XXXXXX XXXXXX XXX XXXXX XXXXXXXXXXXXXXXXXXXXXXX XX XXXXXXXXXXXXXXXXXXXXXXXXXX XX XXX XXXXX XXXXXXXXXXX XXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXX XXXXXX XXXXXXXXXXX XXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XX XXX XXXXXXXXXX XXXXXXXXXXXXXXXXXXX XXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXXXX XXXXXXX XXX XX XXXXXXXX XXXXXXXXXXXX XXXXX XXXXXXXXXXXXXX XXXXXXXXXXXX XXXXXX XXXXXXXXXX XXXXXX XXXXXXXXX XXXX XXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXX XXXX XXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXX XXXXXXXXX XXXXX XX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXX XXXX XXXXXXXXXXX XXXXX XXX XXX XXXXXXXXXX XXX XXXXXXXXXXXX XXXXXXXXXXXXX XXXXX XXX X XXXX XXXXX XX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXX XXXXXXXXXXXX XX XXXXXXXXXXXXXXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXX XXX XXXXXXX XX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXX XXXXXXXXXXXXX XXXXXXXXXX XX XXXXXXXX XXXXX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXX XXXXXX XXXXXXXXXX XXXXXXX XXXXXXX
mit
-3,221,245,733,640,435,000
41.753846
160
0.861411
false
mne-tools/mne-python
mne/tests/test_annotations.py
1
53208
# Authors: Jaakko Leppakangas <[email protected]> # Robert Luke <[email protected]> # # License: BSD 3 clause from collections import OrderedDict from datetime import datetime, timezone from itertools import repeat import sys import os.path as op import pytest from pytest import approx from numpy.testing import (assert_equal, assert_array_equal, assert_array_almost_equal, assert_allclose) import numpy as np import mne from mne import (create_info, read_annotations, annotations_from_events, events_from_annotations) from mne import Epochs, Annotations from mne.utils import (requires_version, catch_logging, requires_pandas) from mne.utils import (assert_and_remove_boundary_annot, _raw_annot, _dt_to_stamp, _stamp_to_dt, check_version) from mne.io import read_raw_fif, RawArray, concatenate_raws from mne.annotations import (_sync_onset, _handle_meas_date, _read_annotations_txt_parse_header) from mne.datasets import testing data_dir = op.join(testing.data_path(download=False), 'MEG', 'sample') fif_fname = op.join(op.dirname(__file__), '..', 'io', 'tests', 'data', 'test_raw.fif') first_samps = pytest.mark.parametrize('first_samp', (0, 10000)) needs_pandas = pytest.mark.skipif( not check_version('pandas'), reason='Needs pandas') # On Windows, datetime.fromtimestamp throws an error for negative times. # We mimic this behavior on non-Windows platforms for ease of testing. class _windows_datetime(datetime): @classmethod def fromtimestamp(cls, timestamp, tzinfo=None): if timestamp < 0: raise OSError('[Errno 22] Invalid argument') return datetime.fromtimestamp(timestamp, tzinfo) @pytest.fixture(scope='function') def windows_like_datetime(monkeypatch): """Ensure datetime.fromtimestamp is Windows-like.""" if not sys.platform.startswith('win'): monkeypatch.setattr('mne.annotations.datetime', _windows_datetime) yield def test_basics(): """Test annotation class.""" raw = read_raw_fif(fif_fname) assert raw.annotations is not None assert len(raw.annotations.onset) == 0 pytest.raises(IOError, read_annotations, fif_fname) onset = np.array(range(10)) duration = np.ones(10) description = np.repeat('test', 10) dt = raw.info['meas_date'] assert isinstance(dt, datetime) stamp = _dt_to_stamp(dt) # Test time shifts. for orig_time in [None, dt, stamp[0], stamp]: annot = Annotations(onset, duration, description, orig_time) if orig_time is None: assert annot.orig_time is None else: assert isinstance(annot.orig_time, datetime) assert annot.orig_time.tzinfo is timezone.utc pytest.raises(ValueError, Annotations, onset, duration, description[:9]) pytest.raises(ValueError, Annotations, [onset, 1], duration, description) pytest.raises(ValueError, Annotations, onset, [duration, 1], description) # Test combining annotations with concatenate_raws raw2 = raw.copy() delta = raw.times[-1] + 1. / raw.info['sfreq'] orig_time = (stamp[0] + stamp[1] * 1e-6 + raw2._first_time) offset = _dt_to_stamp(_handle_meas_date(raw2.info['meas_date'])) offset = offset[0] + offset[1] * 1e-6 offset = orig_time - offset assert_allclose(offset, raw._first_time) annot = Annotations(onset, duration, description, orig_time) assert annot.orig_time is not None assert ' segments' in repr(annot) raw2.set_annotations(annot) assert_allclose(raw2.annotations.onset, onset + offset) assert raw2.annotations is not annot assert raw2.annotations.orig_time is not None concatenate_raws([raw, raw2]) assert_and_remove_boundary_annot(raw) assert_allclose(onset + offset + delta, raw.annotations.onset, rtol=1e-5) assert_array_equal(annot.duration, raw.annotations.duration) assert_array_equal(raw.annotations.description, np.repeat('test', 10)) def test_annot_sanitizing(tmpdir): """Test description sanitizing.""" annot = Annotations([0], [1], ['a;:b']) fname = str(tmpdir.join('custom-annot.fif')) annot.save(fname) annot_read = read_annotations(fname) _assert_annotations_equal(annot, annot_read) # make sure pytest raises error on char-sequence that is not allowed with pytest.raises(ValueError, match='in description not supported'): Annotations([0], [1], ['a{COLON}b']) def test_raw_array_orig_times(): """Test combining with RawArray and orig_times.""" data = np.random.randn(2, 1000) * 10e-12 sfreq = 100. info = create_info(ch_names=['MEG1', 'MEG2'], ch_types=['grad'] * 2, sfreq=sfreq) meas_date = _handle_meas_date(np.pi) info['meas_date'] = meas_date raws = [] for first_samp in [12300, 100, 12]: raw = RawArray(data.copy(), info, first_samp=first_samp) ants = Annotations([1., 2.], [.5, .5], 'x', np.pi + first_samp / sfreq) raw.set_annotations(ants) raws.append(raw) assert_allclose(raws[0].annotations.onset, [124, 125]) raw = RawArray(data.copy(), info) assert not len(raw.annotations) raw.set_annotations(Annotations([1.], [.5], 'x', None)) assert_allclose(raw.annotations.onset, [1.]) raws.append(raw) raw = concatenate_raws(raws, verbose='debug') assert raw.info['meas_date'] == raw.annotations.orig_time == meas_date assert_and_remove_boundary_annot(raw, 3) assert_array_equal(raw.annotations.onset, [124., 125., 134., 135., 144., 145., 154.]) raw.annotations.delete(2) assert_array_equal(raw.annotations.onset, [124., 125., 135., 144., 145., 154.]) raw.annotations.append(5, 1.5, 'y') assert_array_equal(raw.annotations.onset, [5., 124., 125., 135., 144., 145., 154.]) assert_array_equal(raw.annotations.duration, [1.5, .5, .5, .5, .5, .5, .5]) assert_array_equal(raw.annotations.description, ['y', 'x', 'x', 'x', 'x', 'x', 'x']) # These three things should be equivalent stamp = _dt_to_stamp(raw.info['meas_date']) orig_time = _handle_meas_date(stamp) for empty_annot in ( Annotations([], [], [], stamp), Annotations([], [], [], orig_time), Annotations([], [], [], None), None): raw.set_annotations(empty_annot) assert isinstance(raw.annotations, Annotations) assert len(raw.annotations) == 0 assert raw.annotations.orig_time == orig_time def test_crop(tmpdir): """Test cropping with annotations.""" raw = read_raw_fif(fif_fname) events = mne.find_events(raw) onset = events[events[:, 2] == 1, 0] / raw.info['sfreq'] duration = np.full_like(onset, 0.5) description = ['bad %d' % k for k in range(len(onset))] annot = mne.Annotations(onset, duration, description, orig_time=raw.info['meas_date']) raw.set_annotations(annot) split_time = raw.times[-1] / 2. + 2. split_idx = len(onset) // 2 + 1 raw_cropped_left = raw.copy().crop(0., split_time - 1. / raw.info['sfreq']) assert_array_equal(raw_cropped_left.annotations.description, raw.annotations.description[:split_idx]) assert_allclose(raw_cropped_left.annotations.duration, raw.annotations.duration[:split_idx]) assert_allclose(raw_cropped_left.annotations.onset, raw.annotations.onset[:split_idx]) raw_cropped_right = raw.copy().crop(split_time, None) assert_array_equal(raw_cropped_right.annotations.description, raw.annotations.description[split_idx:]) assert_allclose(raw_cropped_right.annotations.duration, raw.annotations.duration[split_idx:]) assert_allclose(raw_cropped_right.annotations.onset, raw.annotations.onset[split_idx:]) raw_concat = mne.concatenate_raws([raw_cropped_left, raw_cropped_right], verbose='debug') assert_allclose(raw_concat.times, raw.times) assert_allclose(raw_concat[:][0], raw[:][0], atol=1e-20) assert_and_remove_boundary_annot(raw_concat) # Ensure we annotations survive round-trip crop->concat assert_array_equal(raw_concat.annotations.description, raw.annotations.description) for attr in ('onset', 'duration'): assert_allclose(getattr(raw_concat.annotations, attr), getattr(raw.annotations, attr), err_msg='Failed for %s:' % (attr,)) raw.set_annotations(None) # undo # Test concatenating annotations with and without orig_time. raw2 = raw.copy() raw.set_annotations(Annotations([45.], [3], 'test', raw.info['meas_date'])) raw2.set_annotations(Annotations([2.], [3], 'BAD', None)) expected_onset = [45., 2. + raw._last_time] raw = concatenate_raws([raw, raw2]) assert_and_remove_boundary_annot(raw) assert_array_almost_equal(raw.annotations.onset, expected_onset, decimal=2) # Test IO tempdir = str(tmpdir) fname = op.join(tempdir, 'test-annot.fif') raw.annotations.save(fname) annot_read = read_annotations(fname) for attr in ('onset', 'duration'): assert_allclose(getattr(annot_read, attr), getattr(raw.annotations, attr)) assert annot_read.orig_time == raw.annotations.orig_time assert_array_equal(annot_read.description, raw.annotations.description) annot = Annotations((), (), ()) annot.save(fname, overwrite=True) pytest.raises(IOError, read_annotations, fif_fname) # none in old raw annot = read_annotations(fname) assert isinstance(annot, Annotations) assert len(annot) == 0 annot.crop() # test if cropping empty annotations doesn't raise an error # Test that empty annotations can be saved with an object fname = op.join(tempdir, 'test_raw.fif') raw.set_annotations(annot) raw.save(fname) raw_read = read_raw_fif(fname) assert isinstance(raw_read.annotations, Annotations) assert len(raw_read.annotations) == 0 raw.set_annotations(None) raw.save(fname, overwrite=True) raw_read = read_raw_fif(fname) assert raw_read.annotations is not None assert len(raw_read.annotations.onset) == 0 @first_samps def test_chunk_duration(first_samp): """Test chunk_duration.""" # create dummy raw raw = RawArray(data=np.empty([10, 10], dtype=np.float64), info=create_info(ch_names=10, sfreq=1.), first_samp=first_samp) raw.info['meas_date'] = _handle_meas_date(0) raw.set_annotations(Annotations(description='foo', onset=[0], duration=[10], orig_time=None)) assert raw.annotations.orig_time == raw.info['meas_date'] assert_allclose(raw.annotations.onset, [first_samp]) # expected_events = [[0, 0, 1], [0, 0, 1], [1, 0, 1], [1, 0, 1], .. # [9, 0, 1], [9, 0, 1]] expected_events = np.atleast_2d(np.repeat(range(10), repeats=2)).T expected_events = np.insert(expected_events, 1, 0, axis=1) expected_events = np.insert(expected_events, 2, 1, axis=1) expected_events[:, 0] += first_samp events, events_id = events_from_annotations(raw, chunk_duration=.5, use_rounding=False) assert_array_equal(events, expected_events) # test chunk durations that do not fit equally in annotation duration expected_events = np.zeros((3, 3)) expected_events[:, -1] = 1 expected_events[:, 0] = np.arange(0, 9, step=3) + first_samp events, events_id = events_from_annotations(raw, chunk_duration=3.) assert_array_equal(events, expected_events) def test_events_from_annotation_orig_time_none(): """Tests events_from_annotation with orig_time None and first_sampe > 0.""" # Create fake data sfreq, duration_s = 100, 10 data = np.random.RandomState(42).randn(1, sfreq * duration_s) info = mne.create_info(ch_names=['EEG1'], ch_types=['eeg'], sfreq=sfreq) raw = mne.io.RawArray(data, info) # Add annotation toward the end onset = [8] duration = [1] description = ['0'] annots = mne.Annotations(onset, duration, description) raw = raw.set_annotations(annots) # Crop start of raw raw.crop(tmin=7) # Extract epochs events, event_ids = mne.events_from_annotations(raw) epochs = mne.Epochs( raw, events, tmin=0, tmax=1, baseline=None, on_missing='warning') # epochs is empty assert_array_equal(epochs.get_data()[0], data[:, 800:901]) def test_crop_more(): """Test more cropping.""" raw = mne.io.read_raw_fif(fif_fname).crop(0, 11).load_data() raw._data[:] = np.random.RandomState(0).randn(*raw._data.shape) onset = np.array([0.47058824, 2.49773765, 6.67873287, 9.15837097]) duration = np.array([0.89592767, 1.13574672, 1.09954739, 0.48868752]) annotations = mne.Annotations(onset, duration, 'BAD') raw.set_annotations(annotations) assert len(raw.annotations) == 4 delta = 1. / raw.info['sfreq'] offset = raw.first_samp * delta raw_concat = mne.concatenate_raws( [raw.copy().crop(0, 4 - delta), raw.copy().crop(4, 8 - delta), raw.copy().crop(8, None)]) assert_allclose(raw_concat.times, raw.times) assert_allclose(raw_concat[:][0], raw[:][0]) assert raw_concat.first_samp == raw.first_samp assert_and_remove_boundary_annot(raw_concat, 2) assert len(raw_concat.annotations) == 4 assert_array_equal(raw_concat.annotations.description, raw.annotations.description) assert_allclose(raw.annotations.duration, duration) assert_allclose(raw_concat.annotations.duration, duration) assert_allclose(raw.annotations.onset, onset + offset) assert_allclose(raw_concat.annotations.onset, onset + offset, atol=1. / raw.info['sfreq']) @testing.requires_testing_data def test_read_brainstorm_annotations(): """Test reading for Brainstorm events file.""" fname = op.join(data_dir, 'events_sample_audvis_raw_bst.mat') annot = read_annotations(fname) assert len(annot) == 238 assert annot.onset.min() > 40 # takes into account first_samp assert np.unique(annot.description).size == 5 @first_samps def test_raw_reject(first_samp): """Test raw data getter with annotation reject.""" sfreq = 100. info = create_info(['a', 'b', 'c', 'd', 'e'], sfreq, ch_types='eeg') raw = RawArray(np.ones((5, 15000)), info, first_samp=first_samp) with pytest.warns(RuntimeWarning, match='outside the data range'): raw.set_annotations(Annotations([2, 100, 105, 148], [2, 8, 5, 8], 'BAD')) data, times = raw.get_data([0, 1, 3, 4], 100, 11200, # 1-112 sec 'omit', return_times=True) bad_times = np.concatenate([np.arange(200, 400), np.arange(10000, 10800), np.arange(10500, 11000)]) expected_times = np.setdiff1d(np.arange(100, 11200), bad_times) / sfreq assert_allclose(times, expected_times) # with orig_time and complete overlap raw = read_raw_fif(fif_fname) raw.set_annotations(Annotations(onset=[1, 4, 5] + raw._first_time, duration=[1, 3, 1], description='BAD', orig_time=raw.info['meas_date'])) t_stop = 18. assert raw.times[-1] > t_stop n_stop = int(round(t_stop * raw.info['sfreq'])) n_drop = int(round(4 * raw.info['sfreq'])) assert len(raw.times) >= n_stop data, times = raw.get_data(range(10), 0, n_stop, 'omit', True) assert data.shape == (10, n_stop - n_drop) assert times[-1] == raw.times[n_stop - 1] assert_array_equal(data[:, -100:], raw[:10, n_stop - 100:n_stop][0]) data, times = raw.get_data(range(10), 0, n_stop, 'NaN', True) assert_array_equal(data.shape, (10, n_stop)) assert times[-1] == raw.times[n_stop - 1] t_1, t_2 = raw.time_as_index([1, 2], use_rounding=True) assert np.isnan(data[:, t_1:t_2]).all() # 1s -2s assert not np.isnan(data[:, :t_1].any()) assert not np.isnan(data[:, t_2:].any()) assert_array_equal(data[:, -100:], raw[:10, n_stop - 100:n_stop][0]) assert_array_equal(raw.get_data(), raw[:][0]) # Test _sync_onset times = [10, -88, 190] onsets = _sync_onset(raw, times) assert_array_almost_equal(onsets, times - raw.first_samp / raw.info['sfreq']) assert_array_almost_equal(times, _sync_onset(raw, onsets, True)) @first_samps def test_annotation_filtering(first_samp): """Test that annotations work properly with filtering.""" # Create data with just a DC component data = np.ones((1, 1000)) info = create_info(1, 1000., 'eeg') raws = [RawArray(data * (ii + 1), info, first_samp=first_samp) for ii in range(4)] kwargs_pass = dict(l_freq=None, h_freq=50., fir_design='firwin') kwargs_stop = dict(l_freq=50., h_freq=None, fir_design='firwin') # lowpass filter, which should not modify the data raws_pass = [raw.copy().filter(**kwargs_pass) for raw in raws] # highpass filter, which should zero it out raws_stop = [raw.copy().filter(**kwargs_stop) for raw in raws] # concat the original and the filtered segments raws_concat = concatenate_raws([raw.copy() for raw in raws]) raws_zero = raws_concat.copy().apply_function(lambda x: x * 0) raws_pass_concat = concatenate_raws(raws_pass) raws_stop_concat = concatenate_raws(raws_stop) # make sure we did something reasonable with our individual-file filtering assert_allclose(raws_concat[0][0], raws_pass_concat[0][0], atol=1e-14) assert_allclose(raws_zero[0][0], raws_stop_concat[0][0], atol=1e-14) # ensure that our Annotations cut up the filtering properly raws_concat_pass = raws_concat.copy().filter(skip_by_annotation='edge', **kwargs_pass) assert_allclose(raws_concat[0][0], raws_concat_pass[0][0], atol=1e-14) raws_concat_stop = raws_concat.copy().filter(skip_by_annotation='edge', **kwargs_stop) assert_allclose(raws_zero[0][0], raws_concat_stop[0][0], atol=1e-14) # one last test: let's cut out a section entirely: # here the 1-3 second window should be skipped raw = raws_concat.copy() raw.annotations.append(1. + raw._first_time, 2., 'foo') with catch_logging() as log: raw.filter(l_freq=50., h_freq=None, fir_design='firwin', skip_by_annotation='foo', verbose='info') log = log.getvalue() assert '2 contiguous segments' in log raw.annotations.append(2. + raw._first_time, 1., 'foo') # shouldn't change with catch_logging() as log: raw.filter(l_freq=50., h_freq=None, fir_design='firwin', skip_by_annotation='foo', verbose='info') log = log.getvalue() assert '2 contiguous segments' in log # our filter will zero out anything not skipped: mask = np.concatenate((np.zeros(1000), np.ones(2000), np.zeros(1000))) expected_data = raws_concat[0][0][0] * mask assert_allclose(raw[0][0][0], expected_data, atol=1e-14) # Let's try another one raw = raws[0].copy() raw.set_annotations(Annotations([0.], [0.5], ['BAD_ACQ_SKIP'])) my_data, times = raw.get_data(reject_by_annotation='omit', return_times=True) assert_allclose(times, raw.times[500:]) assert my_data.shape == (1, 500) raw_filt = raw.copy().filter(skip_by_annotation='bad_acq_skip', **kwargs_stop) expected = data.copy() expected[:, 500:] = 0 assert_allclose(raw_filt[:][0], expected, atol=1e-14) raw = raws[0].copy() raw.set_annotations(Annotations([0.5], [0.5], ['BAD_ACQ_SKIP'])) my_data, times = raw.get_data(reject_by_annotation='omit', return_times=True) assert_allclose(times, raw.times[:500]) assert my_data.shape == (1, 500) raw_filt = raw.copy().filter(skip_by_annotation='bad_acq_skip', **kwargs_stop) expected = data.copy() expected[:, :500] = 0 assert_allclose(raw_filt[:][0], expected, atol=1e-14) @first_samps def test_annotation_omit(first_samp): """Test raw.get_data with annotations.""" data = np.concatenate([np.ones((1, 1000)), 2 * np.ones((1, 1000))], -1) info = create_info(1, 1000., 'eeg') raw = RawArray(data, info, first_samp=first_samp) raw.set_annotations(Annotations([0.5], [1], ['bad'])) expected = raw[0][0] assert_allclose(raw.get_data(reject_by_annotation=None), expected) # nan expected[0, 500:1500] = np.nan assert_allclose(raw.get_data(reject_by_annotation='nan'), expected) got = np.concatenate([raw.get_data(start=start, stop=stop, reject_by_annotation='nan') for start, stop in ((0, 1000), (1000, 2000))], -1) assert_allclose(got, expected) # omit expected = expected[:, np.isfinite(expected[0])] assert_allclose(raw.get_data(reject_by_annotation='omit'), expected) got = np.concatenate([raw.get_data(start=start, stop=stop, reject_by_annotation='omit') for start, stop in ((0, 1000), (1000, 2000))], -1) assert_allclose(got, expected) pytest.raises(ValueError, raw.get_data, reject_by_annotation='foo') def test_annotation_epoching(): """Test that annotations work properly with concatenated edges.""" # Create data with just a DC component data = np.ones((1, 1000)) info = create_info(1, 1000., 'eeg') raw = concatenate_raws([RawArray(data, info) for ii in range(3)]) assert raw.annotations is not None assert len(raw.annotations) == 4 assert np.in1d(raw.annotations.description, ['BAD boundary']).sum() == 2 assert np.in1d(raw.annotations.description, ['EDGE boundary']).sum() == 2 assert_array_equal(raw.annotations.duration, 0.) events = np.array([[a, 0, 1] for a in [0, 500, 1000, 1500, 2000]]) epochs = Epochs(raw, events, tmin=0, tmax=0.999, baseline=None, preload=True) # 1000 samples long assert_equal(len(epochs.drop_log), len(events)) assert_equal(len(epochs), 3) assert_equal([0, 2, 4], epochs.selection) def test_annotation_concat(): """Test if two Annotations objects can be concatenated.""" a = Annotations([1, 2, 3], [5, 5, 8], ["a", "b", "c"]) b = Annotations([11, 12, 13], [1, 2, 2], ["x", "y", "z"]) # test + operator (does not modify a or b) c = a + b assert_array_equal(c.onset, [1, 2, 3, 11, 12, 13]) assert_array_equal(c.duration, [5, 5, 8, 1, 2, 2]) assert_array_equal(c.description, ["a", "b", "c", "x", "y", "z"]) assert_equal(len(a), 3) assert_equal(len(b), 3) assert_equal(len(c), 6) # test += operator (modifies a in place) a += b assert_array_equal(a.onset, [1, 2, 3, 11, 12, 13]) assert_array_equal(a.duration, [5, 5, 8, 1, 2, 2]) assert_array_equal(a.description, ["a", "b", "c", "x", "y", "z"]) assert_equal(len(a), 6) assert_equal(len(b), 3) # test += operator (modifies a in place) b._orig_time = _handle_meas_date(1038942070.7201) with pytest.raises(ValueError, match='orig_time should be the same'): a += b def test_annotations_crop(): """Test basic functionality of annotation crop.""" onset = np.arange(1, 10) duration = np.full_like(onset, 10) description = ["yy"] * onset.shape[0] a = Annotations(onset=onset, duration=duration, description=description, orig_time=0) # cropping window larger than annotations --> do not modify a_ = a.copy().crop(tmin=-10, tmax=42) assert_array_equal(a_.onset, a.onset) assert_array_equal(a_.duration, a.duration) # cropping with left shifted window with pytest.warns(None) as w: a_ = a.copy().crop(tmin=0, tmax=4.2) assert_array_equal(a_.onset, [1., 2., 3., 4.]) assert_allclose(a_.duration, [3.2, 2.2, 1.2, 0.2]) assert len(w) == 0 # cropping with right shifted window with pytest.warns(None) as w: a_ = a.copy().crop(tmin=17.8, tmax=22) assert_array_equal(a_.onset, [17.8, 17.8]) assert_allclose(a_.duration, [0.2, 1.2]) assert len(w) == 0 # cropping with centered small window a_ = a.copy().crop(tmin=11, tmax=12) assert_array_equal(a_.onset, [11, 11, 11, 11, 11, 11, 11, 11, 11]) assert_array_equal(a_.duration, [0, 1, 1, 1, 1, 1, 1, 1, 1]) # cropping with out-of-bounds window with pytest.warns(None) as w: a_ = a.copy().crop(tmin=42, tmax=100) assert_array_equal(a_.onset, []) assert_array_equal(a_.duration, []) assert len(w) == 0 # test error raising with pytest.raises(ValueError, match='tmax should be greater than.*tmin'): a.copy().crop(tmin=42, tmax=0) # test warnings with pytest.warns(RuntimeWarning, match='Omitted .* were outside'): a.copy().crop(tmin=42, tmax=100, emit_warning=True) with pytest.warns(RuntimeWarning, match='Limited .* expanding outside'): a.copy().crop(tmin=0, tmax=12, emit_warning=True) @testing.requires_testing_data def test_events_from_annot_in_raw_objects(): """Test basic functionality of events_fron_annot for raw objects.""" raw = read_raw_fif(fif_fname) events = mne.find_events(raw) event_id = { 'Auditory/Left': 1, 'Auditory/Right': 2, 'Visual/Left': 3, 'Visual/Right': 4, 'Visual/Smiley': 32, 'Motor/Button': 5 } event_map = {v: k for k, v in event_id.items()} annot = Annotations(onset=raw.times[events[:, 0] - raw.first_samp], duration=np.zeros(len(events)), description=[event_map[vv] for vv in events[:, 2]], orig_time=None) raw.set_annotations(annot) events2, event_id2 = \ events_from_annotations(raw, event_id=event_id, regexp=None) assert_array_equal(events, events2) assert_equal(event_id, event_id2) events3, event_id3 = \ events_from_annotations(raw, event_id=None, regexp=None) assert_array_equal(events[:, 0], events3[:, 0]) assert set(event_id.keys()) == set(event_id3.keys()) # ensure that these actually got sorted properly expected_event_id = { desc: idx + 1 for idx, desc in enumerate(sorted(event_id.keys()))} assert event_id3 == expected_event_id first = np.unique(events3[:, 2]) second = np.arange(1, len(event_id) + 1, 1).astype(first.dtype) assert_array_equal(first, second) first = np.unique(list(event_id3.values())) second = np.arange(1, len(event_id) + 1, 1).astype(first.dtype) assert_array_equal(first, second) events4, event_id4 =\ events_from_annotations(raw, event_id=None, regexp='.*Left') expected_event_id4 = {k: v for k, v in event_id.items() if 'Left' in k} assert_equal(event_id4.keys(), expected_event_id4.keys()) expected_events4 = events[(events[:, 2] == 1) | (events[:, 2] == 3)] assert_array_equal(expected_events4[:, 0], events4[:, 0]) events5, event_id5 = \ events_from_annotations(raw, event_id=event_id, regexp='.*Left') expected_event_id5 = {k: v for k, v in event_id.items() if 'Left' in k} assert_equal(event_id5, expected_event_id5) expected_events5 = events[(events[:, 2] == 1) | (events[:, 2] == 3)] assert_array_equal(expected_events5, events5) with pytest.raises(ValueError, match='not find any of the events'): events_from_annotations(raw, regexp='not_there') with pytest.raises(ValueError, match='Invalid type for event_id'): events_from_annotations(raw, event_id='wrong') # concat does not introduce BAD or EDGE raw_concat = concatenate_raws([raw.copy(), raw.copy()]) _, event_id = events_from_annotations(raw_concat) assert isinstance(event_id, dict) assert len(event_id) > 0 for kind in ('BAD', 'EDGE'): assert '%s boundary' % kind in raw_concat.annotations.description for key in event_id.keys(): assert kind not in key # remove all events raw.set_annotations(None) events7, _ = events_from_annotations(raw) assert_array_equal(events7, np.empty((0, 3), dtype=int)) def test_events_from_annot_onset_alingment(): """Test events and annotations onset are the same.""" raw = _raw_annot(meas_date=1, orig_time=1.5) # sec 0 1 2 3 # raw . |--------xxxxxxxxx # annot . |---xx # raw.annot . |--------xx # latency . 0 1 2 # . 0 0 assert raw.annotations.orig_time == _handle_meas_date(1) assert raw.annotations.onset[0] == 1 assert raw.first_samp == 10 event_latencies, event_id = events_from_annotations(raw) assert event_latencies[0, 0] == 10 assert raw.first_samp == event_latencies[0, 0] def _create_annotation_based_on_descr(description, annotation_start_sampl=0, duration=0, orig_time=0): """Create a raw object with annotations from descriptions. The returning raw object contains as many annotations as description given. All starting at `annotation_start_sampl`. """ # create dummy raw raw = RawArray(data=np.empty([10, 10], dtype=np.float64), info=create_info(ch_names=10, sfreq=1000.), first_samp=0) raw.set_meas_date(0) # create dummy annotations based on the descriptions onset = raw.times[annotation_start_sampl] onset_matching_desc = np.full_like(description, onset, dtype=type(onset)) duration_matching_desc = np.full_like(description, duration, dtype=type(duration)) annot = Annotations(description=description, onset=onset_matching_desc, duration=duration_matching_desc, orig_time=orig_time) if duration != 0: with pytest.warns(RuntimeWarning, match='Limited.*expanding outside'): # duration 0.1s is larger than the raw data expand raw.set_annotations(annot) else: raw.set_annotations(annot) # Make sure that set_annotations(annot) works assert all(raw.annotations.onset == onset) if duration != 0: expected_duration = (len(raw.times) / raw.info['sfreq']) - onset else: expected_duration = 0 _duration = raw.annotations.duration[0] assert _duration == approx(expected_duration) assert all(raw.annotations.duration == _duration) assert all(raw.annotations.description == description) return raw def test_event_id_function_default(): """Test[unit_test] for event_id_function default in event_from_annotations. The expected behavior is give numeric label for all those annotations not present in event_id, starting at 1. """ # No event_id given description = ['a', 'b', 'c', 'd', 'e', 'f', 'g'] expected_event_id = dict(zip(description, range(1, 100))) expected_events = np.array([[3, 3, 3, 3, 3, 3, 3], [0, 0, 0, 0, 0, 0, 0], [1, 2, 3, 4, 5, 6, 7]]).T raw = _create_annotation_based_on_descr(description, annotation_start_sampl=3, duration=100) events, event_id = events_from_annotations(raw, event_id=None) assert_array_equal(events, expected_events) assert event_id == expected_event_id def test_event_id_function_using_custom_function(): """Test [unit_test] arbitrary function to create the ids.""" def _constant_id(*args, **kwargs): return 42 description = ['a', 'b', 'c', 'd', 'e', 'f', 'g'] expected_event_id = dict(zip(description, repeat(42))) expected_events = np.repeat([[0, 0, 42]], len(description), axis=0) raw = _create_annotation_based_on_descr(description) events, event_id = events_from_annotations(raw, event_id=_constant_id) assert_array_equal(events, expected_events) assert event_id == expected_event_id # Test for IO with .csv files def _assert_annotations_equal(a, b, tol=0): __tracebackhide__ = True assert_allclose(a.onset, b.onset, rtol=0, atol=tol) assert_allclose(a.duration, b.duration, rtol=0, atol=tol) assert_array_equal(a.description, b.description) assert_array_equal(a.ch_names, b.ch_names) a_orig_time = a.orig_time b_orig_time = b.orig_time assert a_orig_time == b_orig_time _ORIG_TIME = datetime.fromtimestamp(1038942071.7201, timezone.utc) @pytest.fixture(scope='function', params=('ch_names', 'fmt')) def dummy_annotation_file(tmpdir_factory, ch_names, fmt): """Create csv file for testing.""" if fmt == 'csv': content = ("onset,duration,description\n" "2002-12-03 19:01:11.720100,1.0,AA\n" "2002-12-03 19:01:20.720100,2.425,BB") elif fmt == 'txt': content = ("# MNE-Annotations\n" "# orig_time : 2002-12-03 19:01:11.720100\n" "# onset, duration, description\n" "0, 1, AA \n" "9, 2.425, BB") else: assert fmt == 'fif' content = Annotations( [0, 9], [1, 2.425], ['AA', 'BB'], orig_time=_ORIG_TIME) if ch_names: if isinstance(content, Annotations): # this is a bit of a hack but it works content.ch_names[:] = ((), ('MEG0111', 'MEG2563')) else: content = content.splitlines() content[-3] += ',ch_names' content[-2] += ',' content[-1] += ',MEG0111:MEG2563' content = '\n'.join(content) fname = tmpdir_factory.mktemp('data').join(f'annotations-annot.{fmt}') if isinstance(content, str): fname.write(content) else: content.save(fname) return fname @pytest.mark.parametrize('ch_names', (False, True)) @pytest.mark.parametrize('fmt', [ pytest.param('csv', marks=needs_pandas), 'txt', 'fif' ]) def test_io_annotation(dummy_annotation_file, tmpdir, fmt, ch_names): """Test CSV, TXT, and FIF input/output (which support ch_names).""" annot = read_annotations(dummy_annotation_file) assert annot.orig_time == _ORIG_TIME kwargs = dict(orig_time=_ORIG_TIME) if ch_names: kwargs['ch_names'] = ((), ('MEG0111', 'MEG2563')) _assert_annotations_equal( annot, Annotations([0., 9.], [1., 2.425], ['AA', 'BB'], **kwargs), tol=1e-6) # Now test writing fname = tmpdir.join(f'annotations-annot.{fmt}') annot.save(fname) annot2 = read_annotations(fname) _assert_annotations_equal(annot, annot2) # Now without an orig_time annot._orig_time = None annot.save(fname, overwrite=True) annot2 = read_annotations(fname) _assert_annotations_equal(annot, annot2) @requires_version('pandas') def test_broken_csv(tmpdir): """Test broken .csv that does not use timestamps.""" content = ("onset,duration,description\n" "1.,1.0,AA\n" "3.,2.425,BB") fname = tmpdir.join('annotations_broken.csv') fname.write(content) with pytest.warns(RuntimeWarning, match='save your CSV as a TXT'): read_annotations(fname) # Test for IO with .txt files @pytest.fixture(scope='function', params=('ch_names',)) def dummy_annotation_txt_file(tmpdir_factory, ch_names): """Create txt file for testing.""" content = ("3.14, 42, AA \n" "6.28, 48, BB") if ch_names: content = content.splitlines() content[0] = content[0].strip() + ',' content[1] = content[1].strip() + ', MEG0111:MEG2563' content = '\n'.join(content) fname = tmpdir_factory.mktemp('data').join('annotations.txt') fname.write(content) return fname @pytest.mark.parametrize('ch_names', (False, True)) def test_io_annotation_txt(dummy_annotation_txt_file, tmpdir_factory, ch_names): """Test TXT input/output without meas_date.""" annot = read_annotations(str(dummy_annotation_txt_file)) assert annot.orig_time is None kwargs = dict() if ch_names: kwargs['ch_names'] = [(), ('MEG0111', 'MEG2563')] _assert_annotations_equal( annot, Annotations([3.14, 6.28], [42., 48], ['AA', 'BB'], **kwargs)) # Now test writing fname = str(tmpdir_factory.mktemp('data').join('annotations.txt')) annot.save(fname) annot2 = read_annotations(fname) _assert_annotations_equal(annot, annot2) # Now with an orig_time assert annot.orig_time is None annot._orig_time = _handle_meas_date(1038942071.7201) assert annot.orig_time is not None annot.save(fname, overwrite=True) annot2 = read_annotations(fname) assert annot2.orig_time is not None _assert_annotations_equal(annot, annot2) @pytest.mark.parametrize('meas_date, out', [ pytest.param('toto', None, id='invalid string'), pytest.param(None, None, id='None'), pytest.param(42, 42.0, id='Scalar'), pytest.param(3.14, 3.14, id='Float'), pytest.param((3, 140000), 3.14, id='Scalar touple'), pytest.param('2002-12-03 19:01:11.720100', 1038942071.7201, id='valid iso8601 string'), pytest.param('2002-12-03T19:01:11.720100', None, id='invalid iso8601 string')]) def test_handle_meas_date(meas_date, out): """Test meas date formats.""" if out is not None: assert out >= 0 # otherwise it'll break on Windows out = datetime.fromtimestamp(out, timezone.utc) assert _handle_meas_date(meas_date) == out def test_read_annotation_txt_header(tmpdir): """Test TXT orig_time recovery.""" content = ("# A something \n" "# orig_time : 42\n" "# orig_time : 2002-12-03 19:01:11.720100\n" "# orig_time : 42\n" "# C\n" "Done") fname = tmpdir.join('header.txt') fname.write(content) orig_time = _read_annotations_txt_parse_header(fname) want = datetime.fromtimestamp(1038942071.7201, timezone.utc) assert orig_time == want def test_read_annotation_txt_one_segment(tmpdir): """Test empty TXT input/output.""" content = ("# MNE-Annotations\n" "# onset, duration, description\n" "3.14, 42, AA") fname = tmpdir.join('one-annotations.txt') fname.write(content) annot = read_annotations(fname) _assert_annotations_equal(annot, Annotations(3.14, 42, ['AA'])) def test_read_annotation_txt_empty(tmpdir): """Test empty TXT input/output.""" content = ("# MNE-Annotations\n" "# onset, duration, description\n") fname = tmpdir.join('empty-annotations.txt') fname.write(content) annot = read_annotations(fname) _assert_annotations_equal(annot, Annotations([], [], [])) def test_annotations_simple_iteration(): """Test indexing Annotations.""" NUM_ANNOT = 5 EXPECTED_ELEMENTS_TYPE = (np.float64, np.float64, np.str_) EXPECTED_ONSETS = EXPECTED_DURATIONS = [x for x in range(NUM_ANNOT)] EXPECTED_DESCS = [x.__repr__() for x in range(NUM_ANNOT)] annot = Annotations(onset=EXPECTED_ONSETS, duration=EXPECTED_DURATIONS, description=EXPECTED_DESCS, orig_time=None) for ii, elements in enumerate(annot[:2]): assert isinstance(elements, OrderedDict) expected_values = (ii, ii, str(ii)) for elem, expected_type, expected_value in zip(elements.values(), EXPECTED_ELEMENTS_TYPE, expected_values): assert np.isscalar(elem) assert type(elem) == expected_type assert elem == expected_value @requires_version('numpy', '1.12') def test_annotations_slices(): """Test indexing Annotations.""" NUM_ANNOT = 5 EXPECTED_ONSETS = EXPECTED_DURATIONS = [x for x in range(NUM_ANNOT)] EXPECTED_DESCS = [x.__repr__() for x in range(NUM_ANNOT)] annot = Annotations(onset=EXPECTED_ONSETS, duration=EXPECTED_DURATIONS, description=EXPECTED_DESCS, orig_time=None) # Indexing returns a copy. So this has no effect in annot annot[0]['onset'] = 42 annot[0]['duration'] = 3.14 annot[0]['description'] = 'foobar' annot[:1].onset[0] = 42 annot[:1].duration[0] = 3.14 annot[:1].description[0] = 'foobar' # Slicing with single element returns a dictionary for ii in EXPECTED_ONSETS: assert annot[ii] == dict(zip(['onset', 'duration', 'description', 'orig_time'], [ii, ii, str(ii), None])) # Slices should give back Annotations for current in (annot[slice(0, None, 2)], annot[[bool(ii % 2) for ii in range(len(annot))]], annot[:1], annot[[0, 2, 2]], annot[(0, 2, 2)], annot[np.array([0, 2, 2])], annot[1::2], ): assert isinstance(current, Annotations) assert len(current) != len(annot) for bad_ii in [len(EXPECTED_ONSETS), 42, 'foo']: with pytest.raises(IndexError): annot[bad_ii] def test_sorting(): """Test annotation sorting.""" annot = Annotations([10, 20, 30], [1, 2, 3], 'BAD') # assert_array_equal(annot.onset, [0, 5, 10]) annot.append([5, 15, 25, 35], 0.5, 'BAD') onset = list(range(5, 36, 5)) duration = list(annot.duration) assert_array_equal(annot.onset, onset) assert_array_equal(annot.duration, duration) annot.append([10, 10], [0.1, 9], 'BAD') # 0.1 should be before, 9 after want_before = onset.index(10) duration.insert(want_before, 0.1) duration.insert(want_before + 2, 9) onset.insert(want_before, 10) onset.insert(want_before, 10) assert_array_equal(annot.onset, onset) assert_array_equal(annot.duration, duration) def test_date_none(tmpdir): """Test that DATE_NONE is used properly.""" # Regression test for gh-5908 n_chans = 139 n_samps = 20 data = np.random.random_sample((n_chans, n_samps)) ch_names = ['E{}'.format(x) for x in range(n_chans)] ch_types = ['eeg'] * n_chans info = create_info(ch_names=ch_names, ch_types=ch_types, sfreq=2048) assert info['meas_date'] is None raw = RawArray(data=data, info=info) fname = op.join(str(tmpdir), 'test-raw.fif') raw.save(fname) raw_read = read_raw_fif(fname, preload=True) assert raw_read.info['meas_date'] is None def test_negative_meas_dates(windows_like_datetime): """Test meas_date previous to 1970.""" # Regression test for gh-6621 raw = RawArray(data=np.empty((1, 1), dtype=np.float64), info=create_info(ch_names=1, sfreq=1.)) raw.set_meas_date((-908196946, 988669)) raw.set_annotations(Annotations(description='foo', onset=[0], duration=[0], orig_time=None)) events, _ = events_from_annotations(raw) assert events[:, 0] == 0 def test_crop_when_negative_orig_time(windows_like_datetime): """Test cropping with orig_time, tmin and tmax previous to 1970.""" # Regression test for gh-6621 orig_time_stamp = -908196945.011331 # 1941-03-22 11:04:14.988669 annot = Annotations(description='foo', onset=np.arange(0, 0.999, 0.1), duration=[0], orig_time=orig_time_stamp) stamp = _dt_to_stamp(annot.orig_time) assert_allclose(stamp[0] + stamp[1] * 1e-6, orig_time_stamp) t = stamp[0] + stamp[1] * 1e-6 assert t == orig_time_stamp assert len(annot) == 10 # do not raise annot.crop(verbose='debug') assert len(annot) == 10 # Crop with negative tmin, tmax tmin, tmax = [orig_time_stamp + t for t in (0.25, .75)] assert tmin < 0 and tmax < 0 crop_annot = annot.crop(tmin=tmin, tmax=tmax) assert_allclose(crop_annot.onset, [0.3, 0.4, 0.5, 0.6, 0.7]) orig_dt = _stamp_to_dt(stamp) assert crop_annot.orig_time == orig_dt # orig_time does not change def test_allow_nan_durations(): """Deal with "n/a" strings in BIDS events with nan durations.""" raw = RawArray(data=np.empty([2, 10], dtype=np.float64), info=create_info(ch_names=2, sfreq=1.), first_samp=0) raw.set_meas_date(0) ons = [1, 2., 15., 17.] dus = [np.nan, 1., 0.5, np.nan] descriptions = ['A'] * 4 onsets = np.asarray(ons, dtype=float) durations = np.asarray(dus, dtype=float) annot = mne.Annotations(onset=onsets, duration=durations, description=descriptions) with pytest.warns(RuntimeWarning, match='Omitted 2 annotation'): raw.set_annotations(annot) @testing.requires_testing_data def test_annotations_from_events(): """Test events to annotations conversion.""" raw = read_raw_fif(fif_fname) events = mne.find_events(raw) # 1. Automatic event description # ------------------------------------------------------------------------- annots = annotations_from_events(events, raw.info['sfreq'], first_samp=raw.first_samp, orig_time=None) assert len(annots) == events.shape[0] # Convert back to events raw.set_annotations(annots) events_out, _ = events_from_annotations(raw, event_id=int) assert_array_equal(events, events_out) # 2. Explicit event mapping # ------------------------------------------------------------------------- event_desc = {1: 'one', 2: 'two', 3: 'three', 32: None} annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=event_desc, first_samp=raw.first_samp, orig_time=None) assert np.all([a in ['one', 'two', 'three'] for a in annots.description]) assert len(annots) == events[events[:, 2] <= 3].shape[0] # 3. Pass list # ------------------------------------------------------------------------- event_desc = [1, 2, 3] annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=event_desc, first_samp=raw.first_samp, orig_time=None) assert np.all([a in ['1', '2', '3'] for a in annots.description]) assert len(annots) == events[events[:, 2] <= 3].shape[0] # 4. Try passing callable # ------------------------------------------------------------------------- event_desc = lambda d: 'event{}'.format(d) # noqa:E731 annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=event_desc, first_samp=raw.first_samp, orig_time=None) assert np.all(['event' in a for a in annots.description]) assert len(annots) == events.shape[0] # 5. Pass numpy array # ------------------------------------------------------------------------- event_desc = np.array([[1, 2, 3], [1, 2, 3]]) with pytest.raises(ValueError, match='event_desc must be 1D'): annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=event_desc, first_samp=raw.first_samp, orig_time=None) with pytest.raises(ValueError, match='Invalid type for event_desc'): annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=1, first_samp=raw.first_samp, orig_time=None) event_desc = np.array([1, 2, 3]) annots = annotations_from_events(events, sfreq=raw.info['sfreq'], event_desc=event_desc, first_samp=raw.first_samp, orig_time=None) assert np.all([a in ['1', '2', '3'] for a in annots.description]) assert len(annots) == events[events[:, 2] <= 3].shape[0] def test_repr(): """Test repr of Annotations.""" # short annotation repr (< 79 characters) r = repr(Annotations(range(3), [0] * 3, list("abc"))) assert r == '<Annotations | 3 segments: a (1), b (1), c (1)>' # long annotation repr (> 79 characters, will be shortened) r = repr(Annotations(range(14), [0] * 14, list("abcdefghijklmn"))) assert r == ('<Annotations | 14 segments: a (1), b (1), c (1), d (1), ' 'e (1), f (1), g ...>') # empty Annotations r = repr(Annotations([], [], [])) assert r == '<Annotations | 0 segments>' @requires_pandas def test_annotation_to_data_frame(): """Test annotation class to data frame conversion.""" onset = np.arange(1, 10) durations = np.full_like(onset, [4, 5, 6, 4, 5, 6, 4, 5, 6]) description = ["yy"] * onset.shape[0] a = Annotations(onset=onset, duration=durations, description=description, orig_time=0) df = a.to_data_frame() for col in ['onset', 'duration', 'description']: assert col in df.columns assert df.description[0] == 'yy' assert (df.onset[1] - df.onset[0]).seconds == 1 assert df.groupby('description').count().onset['yy'] == 9 def test_annotation_ch_names(): """Test annotation ch_names updating and pruning.""" info = create_info(10, 1000., 'eeg') raw = RawArray(np.zeros((10, 1000)), info) onset = [0.1, 0.3, 0.6] duration = [0.05, 0.1, 0.2] description = ['first', 'second', 'third'] ch_names = [[], raw.ch_names[4:6], raw.ch_names[5:7]] annot = Annotations(onset, duration, description, ch_names=ch_names) raw.set_annotations(annot) # renaming rename = {name: name + 'new' for name in raw.ch_names} raw_2 = raw.copy().rename_channels(rename) for ch_rename, ch in zip(raw_2.annotations.ch_names, annot.ch_names): assert all(name in raw_2.ch_names for name in ch_rename) assert all(name in raw.ch_names for name in ch) assert not any(name in raw.ch_names for name in ch_rename) assert not any(name in raw_2.ch_names for name in ch) raw_2.rename_channels({val: key for key, val in rename.items()}) _assert_annotations_equal(raw.annotations, raw_2.annotations) # dropping raw_2.drop_channels(raw.ch_names[5:]) annot_pruned = raw_2.annotations assert len(raw_2.annotations) == 2 # dropped the last one assert raw_2.annotations.ch_names[1] == tuple(raw.ch_names[4:5]) for ch_drop in raw_2.annotations.ch_names: assert all(name in raw_2.ch_names for name in ch_drop) with pytest.raises(ValueError, match='channel name in annotations missin'): raw_2.set_annotations(annot) with pytest.warns(RuntimeWarning, match='channel name in annotations mis'): raw_2.set_annotations(annot, on_missing='warn') assert raw_2.annotations is not annot_pruned _assert_annotations_equal(raw_2.annotations, annot_pruned) def test_annotation_duration_setting(): """Test annotation duration setting works.""" a = Annotations([1, 2, 3], [5, 5, 8], ["a", "b", "c"]) assert len(a) == 3 assert a.duration[0] == 5 assert a.duration[2] == 8 a.set_durations({"a": 3}) assert a.duration[0] == 3 assert a.duration[2] == 8 a.set_durations({"a": 313, "c": 18}) assert a.duration[0] == 313 assert a.duration[2] == 18 a.set_durations({"a": 1, "b": 13}) assert a.duration[0] == 1 assert a.duration[1] == 13 a = Annotations([1, 2, 3], [5, 5, 8], ["a", "b", "c"]) assert len(a) == 3 assert a.duration[0] == 5 assert a.duration[2] == 8 a.set_durations(7.2) assert a.duration[0] == 7.2 assert a.duration[2] == 7.2 with pytest.raises(ValueError, match="mapping missing from data"): a.set_durations({"aaa": 2.2}) with pytest.raises(ValueError, match="<class 'set'> was provided"): a.set_durations({"aaa", 2.2})
bsd-3-clause
-2,619,110,020,765,893,600
39.431611
79
0.596583
false
andrewsosa/hackfsu_com
api/api/views/hackathon/get/stats.py
2
4468
""" Get public statistics for current hackathon """ from django import forms from django.http.request import HttpRequest from hackfsu_com.views.generic import ApiView from hackfsu_com.util import acl from api.models import Hackathon, HackerInfo, MentorInfo, JudgeInfo, OrganizerInfo, AttendeeStatus from django.utils import timezone class ResponseForm(forms.Form): hackathon_name = forms.CharField() hackathon_start = forms.DateField() hackathon_end = forms.DateField() hackers_registered = forms.IntegerField() hackers_approved = forms.IntegerField() hackers_rsvp = forms.IntegerField() hackers_checked_in = forms.IntegerField() mentors_registered = forms.IntegerField() mentors_approved = forms.IntegerField() mentors_rsvp = forms.IntegerField() mentors_checked_in = forms.IntegerField() judges_registered = forms.IntegerField() judges_approved = forms.IntegerField() judges_rsvp = forms.IntegerField() judges_checked_in = forms.IntegerField() organizers_registered = forms.IntegerField() organizers_approved = forms.IntegerField() organizers_rsvp = forms.IntegerField() organizers_checked_in = forms.IntegerField() class StatsView(ApiView): response_form_class = ResponseForm http_method_names = ['get'] access_manager = acl.AccessManager(acl_accept=[acl.group_user]) def work(self, request, req: dict, res: dict): ch = Hackathon.objects.current() res['hackathon_name'] = ch.name res['hackathon_start'] = ch.start_date res['hackathon_end'] = ch.end_date if (timezone.now().date() - ch.start_date).days >= 0 or \ OrganizerInfo.objects.filter(hackathon=ch, user=request.user, approved=True).exists(): res['hackers_registered'] = HackerInfo.objects.filter(hackathon=ch).count() res['hackers_approved'] = HackerInfo.objects.filter(hackathon=ch, approved=True).count() res['hackers_rsvp'] = \ HackerInfo.objects.filter(hackathon=ch, approved=True, attendee_status__rsvp_result=1).count() res['hackers_checked_in'] = HackerInfo.objects.filter( hackathon=ch, approved=True, attendee_status__checked_in_at__isnull=False).count() res['mentors_registered'] = MentorInfo.objects.filter(hackathon=ch).count() res['mentors_approved'] = MentorInfo.objects.filter(hackathon=ch, approved=True).count() res['mentors_rsvp'] = \ MentorInfo.objects.filter(hackathon=ch, approved=True, attendee_status__rsvp_result=1).count() res['mentors_checked_in'] = MentorInfo.objects.filter( hackathon=ch, approved=True, attendee_status__checked_in_at__isnull=False).count() res['judges_registered'] = JudgeInfo.objects.filter(hackathon=ch).count() res['judges_approved'] = JudgeInfo.objects.filter(hackathon=ch, approved=True).count() res['judges_rsvp'] = \ JudgeInfo.objects.filter(hackathon=ch, approved=True, attendee_status__rsvp_result=1).count() res['judges_checked_in'] = JudgeInfo.objects.filter( hackathon=ch, approved=True, attendee_status__checked_in_at__isnull=False).count() res['organizers_registered'] = OrganizerInfo.objects.filter(hackathon=ch).count() res['organizers_approved'] = OrganizerInfo.objects.filter(hackathon=ch, approved=True).count() res['organizers_rsvp'] = \ OrganizerInfo.objects.filter(hackathon=ch, approved=True, attendee_status__rsvp_result=1).count() res['organizers_checked_in'] = OrganizerInfo.objects.filter( hackathon=ch, approved=True, attendee_status__checked_in_at__isnull=False).count() else: res['hackers_registered'] = -1 res['hackers_approved'] = -1 res['hackers_rsvp'] = -1 res['hackers_checked_in'] = -1 res['mentors_registered'] = -1 res['mentors_approved'] = -1 res['mentors_rsvp'] = -1 res['mentors_checked_in'] = -1 res['judges_registered'] = -1 res['judges_approved'] = -1 res['judges_rsvp'] = -1 res['judges_checked_in'] = -1 res['organizers_registered'] = -1 res['organizers_approved'] = -1 res['organizers_rsvp'] = -1 res['organizers_checked_in'] = -1
apache-2.0
7,364,637,496,740,350,000
47.043011
113
0.643912
false
sevein/archivematica
src/dashboard/src/contrib/utils.py
1
2104
# This file is part of Archivematica. # # Copyright 2010-2013 Artefactual Systems Inc. <http://artefactual.com> # # Archivematica is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Archivematica is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Archivematica. If not, see <http://www.gnu.org/licenses/>. import os def get_directory_size(path='.'): total_size = 0 for dirpath, dirnames, filenames in os.walk(path): for f in filenames: fp = os.path.join(dirpath, f) total_size += os.path.getsize(fp) return total_size def get_directory_name(directory, default=None): """ Attempts to extract a directory name given a transfer or SIP path. Expected format: %sharedPath%watchedDirectories/workFlowDecisions/createDip/ImagesSIP-69826e50-87a2-4370-b7bd-406fc8aad94f/ Given this example, this function would return 'ImagesSIP'. If the optional `default` keyword argument is passed in, the provided value will be used if no name can be extracted. """ import re try: return re.search(r'^.*/(?P<directory>.*)-[\w]{8}(-[\w]{4}){3}-[\w]{12}[/]{0,1}$', directory).group('directory') except: pass try: return re.search(r'^.*/(?P<directory>.*)/$', directory).group('directory') except: pass if directory: return directory else: return default def get_directory_name_from_job(jobs): try: job = jobs[0] # No jobs yet, e.g. not started; there will be no directory name yet except IndexError: return "(Unnamed)" return get_directory_name(job.directory, default=job.sipuuid)
agpl-3.0
4,646,505,071,088,285,000
32.396825
123
0.681559
false
basak/netkeyscript
netkeyscript-send.py
1
2002
#!/usr/bin/python # Copyright 2012 Robie Basak # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL # THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR # OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, # ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR # OTHER DEALINGS IN THE SOFTWARE. NETKEYSCRIPT_PROTO_PASSPHRASE = 0 import argparse import struct import sys from scapy.all import ( Ether, IPv6, UDP, sendp ) def send(dst, sport, dport, payload, iface): ether = Ether() ip = IPv6(dst=dst) udp = UDP(sport=sport, dport=dport) sendp(ether / ip / udp / payload, iface=iface) def main(): parser = argparse.ArgumentParser() parser.add_argument('--iface', default='eth0') parser.add_argument('--sport', default=30621, type=int) parser.add_argument('--dport', default=30621, type=int) parser.add_argument('--dest', default='ff02::1') args = parser.parse_args() payload_command = struct.pack('b', NETKEYSCRIPT_PROTO_PASSPHRASE) payload = payload_command + sys.stdin.read() send(dst=args.dest, sport=args.sport, dport=args.dport, payload=payload, iface=args.iface) if __name__ == '__main__': main()
mit
-4,701,440,638,697,542,000
32.932203
78
0.720779
false
algolia/algoliasearch-django
tests/settings.py
1
2390
""" Django settings for core project. Generated by 'django-admin startproject' using Django 1.8.2. For more information on this file, see https://docs.djangoproject.com/en/1.8/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.8/ref/settings/ """ import os BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) SECRET_KEY = 'MillisecondsMatter' DEBUG = False # Application definition INSTALLED_APPS = ( 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'algoliasearch_django', 'tests' ) MIDDLEWARE = [ 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.auth.middleware.SessionAuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'django.middleware.security.SecurityMiddleware', ] TEMPLATES = [ { "BACKEND": "django.template.backends.django.DjangoTemplates", "DIRS": [], "APP_DIRS": True, "OPTIONS": { "context_processors": [ "django.template.context_processors.debug", "django.template.context_processors.request", "django.contrib.auth.context_processors.auth", "django.contrib.messages.context_processors.messages", ], }, }, ] ROOT_URLCONF = 'tests.urls' # Database DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } # Internationalization LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True def safe_index_name(name): if 'TRAVIS' not in os.environ: return name job = os.environ['TRAVIS_JOB_NUMBER'] return '{}_travis-{}'.format(name, job) # AlgoliaSearch settings ALGOLIA = { 'APPLICATION_ID': os.getenv('ALGOLIA_APPLICATION_ID'), 'API_KEY': os.getenv('ALGOLIA_API_KEY'), 'INDEX_PREFIX': 'test', 'INDEX_SUFFIX': safe_index_name('django'), 'RAISE_EXCEPTIONS': True }
mit
176,922,860,896,906,560
25.555556
70
0.667364
false
all-of-us/raw-data-repository
rdr_service/tools/update_release_tracker.py
1
2190
#!/usr/bin/env python """Updates JIRA release notes when deploying to an environment. This requires the JIRA_API_USER_PASSWORD and JIRA_API_USER_NAME environment variables to be set, and flags for version and instance to be provided. """ import logging import os import sys import jira from rdr_service.main_util import configure_logging, get_parser _JIRA_INSTANCE_URL = "https://precisionmedicineinitiative.atlassian.net/" # Release tickets are moved from our usual project, DA, to the PD project # for change approval, so for stable/prod releases look for tickets there. _JIRA_PROJECT_ID = "PD" def _connect_to_jira(jira_username, jira_password): return jira.JIRA(_JIRA_INSTANCE_URL, basic_auth=(jira_username, jira_password)) def main(args): jira_username = os.getenv("JIRA_API_USER_NAME") jira_password = os.getenv("JIRA_API_USER_PASSWORD") if not jira_username or not jira_password: logging.error("JIRA_API_USER_NAME and JIRA_API_USER_PASSWORD variables must be set. Exiting.") sys.exit(-1) jira_connection = _connect_to_jira(jira_username, jira_password) summary = "Release tracker for %s" % args.version issues = jira_connection.search_issues( 'project = "%s" AND summary ~ "%s" ORDER BY created DESC' % (_JIRA_PROJECT_ID, summary) ) if issues: if len(issues) > 1: logging.warning( "Found multiple release tracker matches, using newest. %s", ", ".join("[%s] %s" % (issue.key, issue.fields().summary) for issue in issues), ) issue = issues[0] jira_connection.add_comment(issue, args.comment) logging.info("Updated issue %s", issue.key) sys.exit(0) else: logging.error("No issue found with summary %r in project %r; exiting.", summary, _JIRA_PROJECT_ID) sys.exit(-1) if __name__ == "__main__": configure_logging() parser = get_parser() parser.add_argument("--version", help="The version of the app being deployed (e.g. v0-1-rc21", required=True) parser.add_argument("--comment", type=str, help="The comment to add to the issue", required=True) main(parser.parse_args())
bsd-3-clause
-2,994,243,296,558,638,600
36.118644
113
0.66895
false
mark-r-g/hydrus
tests/test_model.py
1
1242
# Mark Gatheman <[email protected]> # # This file is part of Hydrus. # # Hydrus is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Hydrus is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Hydrus. If not, see <http://www.gnu.org/licenses/>. import numpy as np from numpy.testing import assert_approx_equal from scipy.stats import norm from hypothesis import given from hydrus.model import Lvm from tests import strat_1d, strat_pos_1d @given(strat_1d, strat_1d, strat_1d, strat_1d, strat_pos_1d, strat_1d) def test_preds_ll(alpha, mu, gamma, err, num, w): current_impl = Lvm.preds_ll(alpha, mu, gamma, err, num, w) simple_impl = np.nansum(w * norm.logpdf(num, mu+gamma*alpha, err)) simple_impl += np.sum(norm.logpdf(alpha)) assert_approx_equal(current_impl, simple_impl)
gpl-3.0
-6,105,238,677,537,839,000
37.8125
70
0.741546
false
google/earthengine-community
samples/python/apidocs/ee-dictionary-select.py
1
1194
# Copyright 2021 The Google Earth Engine Community Authors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # [START earthengine__apidocs__ee_dictionary_select] # A dictionary (e.g. results of ee.Image.reduceRegion of an S2 image). dic = ee.Dictionary({ 'B1': 182, 'B2': 219, 'B3': 443 }) print('Select keys by name:', dic.select(['B1', 'B2']).getInfo()) print('Select keys by regex:', dic.select(['B[1-2]']).getInfo()) dic_select = dic.select(**{'selectors': ['B1', 'B2', 'Region'], 'ignoreMissing': True}) print('Set ignoreMissing as true to avoid an unmatched key error:', dic_select.getInfo()) # [END earthengine__apidocs__ee_dictionary_select]
apache-2.0
-7,978,989,095,025,099,000
38.8
74
0.697655
false
AdaptivePELE/AdaptivePELE
AdaptivePELE/analysis/crossValidationClustering.py
1
3872
from __future__ import absolute_import, division, print_function, unicode_literals import os import argparse import matplotlib.pyplot as plt import numpy as np import pyemma.coordinates as coor import pyemma.msm as msm from AdaptivePELE.freeEnergies import cluster plt.switch_backend("pdf") plt.style.use("ggplot") def parse_arguments(): """ Create command-line interface """ desc = "Plot information related to an MSM" parser = argparse.ArgumentParser(description=desc) parser.add_argument("-l", "--lagtimes", type=int, nargs="*", help="Lagtimes to analyse") parser.add_argument("-c", "--clusters", type=int, nargs="*", help="Number of clusters to analyse") parser.add_argument("-m", type=int, default=6, help="Number of eigenvalues to sum in the GMRQ") parser.add_argument("--tica", action="store_true", help="Whether to use TICA before clustering") parser.add_argument("--tica_lag", type=int, default=30, help="Lagtime for the TICA estimation") parser.add_argument("--out_path", type=str, default="", help="Path to store the output") args = parser.parse_args() return args.lagtimes, args.m, args.tica, args.tica_lag, args.out_path, args.clusters def main(lagtimes, clusters, m, tica_lag, tica, output_path): trajectoryFolder = "allTrajs" trajectoryBasename = "traj*" stride = 1 if output_path and not os.path.exists(output_path): os.makedirs(output_path) scores_path = os.path.join(output_path, "scores") if not os.path.exists(scores_path): os.makedirs(scores_path) data, _ = cluster.loadTrajFiles(trajectoryFolder, trajectoryBasename) if tica: tica_obj = coor.tica(data, lag=tica_lag, var_cutoff=0.9, kinetic_map=True) print('TICA dimension ', tica_obj.dimension()) data = tica_obj.get_output() for tau in lagtimes: scores = [] scores_cv = [] print("Estimating MSM with %d lagtime" % tau) for k in clusters: print("Calculating scores with %d clusters" % k) # cluster data cl = coor.cluster_kmeans(data=data, k=k, max_iter=500, stride=stride) try: MSM = msm.estimate_markov_model(cl.dtrajs, tau) print("MSM estimated on %d states" % MSM.nstates) except Exception: print("Estimation error in %d clusters, %d lagtime" % (k, tau)) scores.append(0) scores_cv.append(np.array([0, 0, 0, 0, 0])) continue try: scores.append(MSM.score(MSM.dtrajs_full, score_k=m)) except Exception: print("Estimation error in %d clusters, %d lagtime" % (k, tau)) scores.append(0) scores_cv.append(np.array([0, 0, 0, 0, 0])) continue try: scores_cv.append(MSM.score_cv(MSM.dtrajs_full, score_k=m, n=5)) except Exception: print("Estimation error in %d clusters, %d lagtime" % (k, tau)) scores_cv.append(np.array([0, 0, 0, 0, 0])) np.save(os.path.join(scores_path, "scores_lag_%d.npy" % tau), scores) np.save(os.path.join(scores_path, "scores_cv_lag_%d.npy" % tau), scores_cv) mean_scores = [sc.mean() for sc in scores_cv] std_scores = [sc.std() for sc in scores_cv] plt.figure() plt.plot(clusters, scores, label="Training") plt.errorbar(clusters, mean_scores, yerr=std_scores, fmt='k', label="Testing") plt.xlabel("Number of states") plt.ylabel("Score") plt.legend() plt.savefig(os.path.join(output_path, "scores_cv_lag_%d.png" % tau)) if __name__ == "__main__": lags, GMRQ, use_tica, lag_tica, out_path, cluster_list = parse_arguments() main(lags, cluster_list, GMRQ, lag_tica, use_tica, out_path)
mit
-8,509,077,382,591,035,000
44.552941
102
0.612603
false
GoogleCloudPlatform/sap-deployment-automation
third_party/github.com/ansible/awx/awx/main/tests/functional/api/test_unified_job_template.py
1
2858
import pytest from awx.api.versioning import reverse from awx.main import models @pytest.mark.django_db def test_aliased_forward_reverse_field_searches(instance, options, get, admin): url = reverse('api:unified_job_template_list') response = options(url, None, admin) assert 'job_template__search' in response.data['related_search_fields'] get(reverse("api:unified_job_template_list") + "?job_template__search=anything", user=admin, expect=200) @pytest.mark.django_db @pytest.mark.parametrize('model', ( 'Project', 'JobTemplate', 'WorkflowJobTemplate' )) class TestUnifiedOrganization: def data_for_model(self, model, orm_style=False): data = { 'name': 'foo', 'organization': None } if model == 'JobTemplate': proj = models.Project.objects.create( name="test-proj", playbook_files=['helloworld.yml'] ) if orm_style: data['project_id'] = proj.id else: data['project'] = proj.id data['playbook'] = 'helloworld.yml' data['ask_inventory_on_launch'] = True return data def test_organization_blank_on_edit_of_orphan(self, model, admin_user, patch): cls = getattr(models, model) data = self.data_for_model(model, orm_style=True) obj = cls.objects.create(**data) patch( url=obj.get_absolute_url(), data={'name': 'foooooo'}, user=admin_user, expect=200 ) obj.refresh_from_db() assert obj.name == 'foooooo' def test_organization_blank_on_edit_of_orphan_as_nonsuperuser(self, model, rando, patch): """Test case reflects historical bug where ordinary users got weird error message when editing an orphaned project """ cls = getattr(models, model) data = self.data_for_model(model, orm_style=True) obj = cls.objects.create(**data) if model == 'JobTemplate': obj.project.admin_role.members.add(rando) obj.admin_role.members.add(rando) patch( url=obj.get_absolute_url(), data={'name': 'foooooo'}, user=rando, expect=200 ) obj.refresh_from_db() assert obj.name == 'foooooo' def test_organization_blank_on_edit_of_normal(self, model, admin_user, patch, organization): cls = getattr(models, model) data = self.data_for_model(model, orm_style=True) data['organization'] = organization obj = cls.objects.create(**data) patch( url=obj.get_absolute_url(), data={'name': 'foooooo'}, user=admin_user, expect=200 ) obj.refresh_from_db() assert obj.name == 'foooooo'
apache-2.0
-8,665,169,606,486,610,000
32.623529
108
0.580126
false
cloew/KaoJson
kao_json/Test/test_conversion_config.py
1
1361
from .class_hierarchy import A, AChild, AGrandchild from ..conversion_config import ConversionConfig import unittest class find(unittest.TestCase): """ Test cases of find """ def test_matchingClassFound(self): """ Test that a matching class is found properly """ expected = "Dummy Config..." config = ConversionConfig([{A:expected}]) actual = config.find(A) self.assertEqual(expected, actual) def test_descendantClassFound(self): """ Test that a matching descendant class is found properly """ expected = "Dummy Config..." config = ConversionConfig([{A:expected}]) actual = config.find(AGrandchild) self.assertEqual(expected, actual) def test_noMatchFound(self): """ Test that when no match is found, None is returned """ config = ConversionConfig([]) self.assertIsNone(config.find(A)) class newConverter(unittest.TestCase): """ Test cases of newConverter """ def test_converterBuilt(self): """ Test that the Converter was built properly """ value = "Dummy Value..." config = ConversionConfig([]) converter = config.newConverter(value) self.assertEqual(converter.value, value) self.assertEqual(converter.config, config)
mit
-7,443,373,079,876,724,000
33.05
71
0.62601
false
agx/git-buildpackage
gbp/deb/rollbackgit.py
1
5051
# vim: set fileencoding=utf-8 : # # (C) 2018 Guido Günther <[email protected]> """A git repository for Debian packages that can roll back operations""" from .. import log from .. git import GitRepositoryError from . git import DebianGitRepository class RollbackError(GitRepositoryError): """ Error raised if the rollback failed """ def __init__(self, errors): self.msg = "Automatic rollback failed" super(RollbackError, self).__init__(self.msg) self.errors = errors def __str__(self): return "%s %s" % (self.msg, self.errors) class RollbackDebianGitRepository(DebianGitRepository): """ Like a DebianGitRepository but can also perform rollbacks and knows about some of the inner workings upstream vcs_tag, … """ def __init__(self, *args, **kwargs): self.rollbacks = [] self.rollback_errors = [] DebianGitRepository.__init__(self, *args, **kwargs) def has_rollbacks(self): return len(self.rollbacks) > 0 def rrr(self, refname, action, reftype): """ Remember ref for rollback @param refname: ref to roll back @param action: the rollback action (delete, reset, ...) @param reftype: the reference type (tag, branch, ...) """ sha = None if action == 'reset': try: sha = self.rev_parse(refname) except GitRepositoryError as err: log.warn("Failed to rev-parse '%s': %s" % (refname, err)) elif action == 'delete': pass elif action == 'abortmerge': pass else: raise GitRepositoryError("Unknown action '%s' for %s '%s'" % (action, reftype, refname)) self.rollbacks.append((refname, reftype, action, sha)) def rrr_branch(self, branchname, action='reset-or-delete'): if action == 'reset-or-delete': if self.has_branch(branchname): return self.rrr(branchname, 'reset', 'branch') else: return self.rrr(branchname, 'delete', 'branch') else: return self.rrr(branchname, action, 'branch') def rrr_tag(self, tagname, action='delete'): return self.rrr(tagname, action, 'tag') def rrr_merge(self, commit, action='abortmerge'): return self.rrr(commit, action, 'commit') def rollback(self): """ Perform a complete rollback Try to roll back as much as possible and remember what failed. """ for (name, reftype, action, sha) in self.rollbacks: try: if action == 'delete': log.info("Rolling back %s '%s' by deleting it" % (reftype, name)) if reftype == 'tag': self.delete_tag(name) elif reftype == 'branch': self.delete_branch(name) else: raise GitRepositoryError("Don't know how to delete %s '%s'" % (reftype, name)) elif action == 'reset' and reftype == 'branch': log.info('Rolling back branch %s by resetting it to %s' % (name, sha)) self.update_ref("refs/heads/%s" % name, sha, msg="gbp import-orig: failure rollback of %s" % name) elif action == 'abortmerge': if self.is_in_merge(): log.info('Rolling back failed merge of %s' % name) self.abort_merge() else: log.info("Nothing to rollback for merge of '%s'" % name) else: raise GitRepositoryError("Don't know how to %s %s '%s'" % (action, reftype, name)) except GitRepositoryError as e: self.rollback_errors.append((name, reftype, action, sha, e)) if self.rollback_errors: raise RollbackError(self.rollback_errors) # Wrapped methods for rollbacks def create_tag(self, *args, **kwargs): name = kwargs['name'] ret = super(RollbackDebianGitRepository, self).create_tag(*args, **kwargs) self.rrr_tag(name) return ret def commit_dir(self, *args, **kwargs): import_branch = kwargs['branch'] self.rrr_branch(import_branch) return super(RollbackDebianGitRepository, self).commit_dir(*args, **kwargs) def create_branch(self, *args, **kwargs): branch = kwargs['branch'] ret = super(RollbackDebianGitRepository, self).create_branch(*args, **kwargs) self.rrr_branch(branch, 'delete') return ret def merge(self, *args, **kwargs): commit = args[0] if args else kwargs['commit'] try: return super(RollbackDebianGitRepository, self).merge(*args, **kwargs) except GitRepositoryError: # Only cleanup in the error case to undo working copy # changes. Resetting the refs handles the other cases. self.rrr_merge(commit) raise
gpl-2.0
3,977,356,638,823,929,300
36.671642
118
0.565571
false
conversationai/wikidetox
experimental/conversation_go_awry/feature_extraction/utils/write_lexicons.py
1
1163
import json lexicon = { 'pron_me': ['i', "i'd", "i'll", "i'm", "i've", 'id', 'im', 'ive', 'me', 'mine', 'my', 'myself'], 'pron_we': ["let's", 'lets', 'our', 'ours', 'ourselves', 'us', 'we', "we'd", "we'll", "we're", "we've", 'weve'], 'pron_you': ["y'all", 'yall', 'you', "you'd", "you'll", "you're", "you've", 'youd', 'youll', 'your', 'youre', 'yours', 'youve'], 'pron_3rd': ['he', "he'd", "he's", 'hed', 'her', 'hers', 'herself', 'hes', 'him', 'himself', 'his', 'she', "she'd", "she'll", "she's", 'shes'], 'pron_3rd_plural': ['their', 'them', 'themselves', 'they', "they'd", "they'll", "they've", 'theyd', 'theyll', 'theyve', "they're", "theyre"] } lexicon['positive'] = [] with open('liu-positive-words.txt') as f: for line in f: lexicon['positive'].append(line.strip()) lexicon['negative'] = [] with open('liu-negative-words.txt', encoding='ISO-8859-1') as f: for line in f: lexicon['negative'].append(line.strip()) with open('lexicons', 'w') as w: json.dump(lexicon, w)
apache-2.0
-5,095,081,116,083,255,000
40.535714
75
0.468616
false
hcrlab/access_teleop
limb_manipulation/src/limb_pbd_server_with_moveit_commander.py
1
43291
#! /usr/bin/env python import rospy import math from pprint import pprint import numpy as np import fetch_api from std_msgs.msg import String, Header, ColorRGBA, Bool from std_srvs.srv import Empty from image_geometry import PinholeCameraModel import tf import tf.transformations as tft # from tf import TransformBroadcaster from geometry_msgs.msg import Pose, PoseStamped, Quaternion, Point, Vector3 from interactive_markers.interactive_marker_server import InteractiveMarkerServer from visualization_msgs.msg import Marker, MarkerArray, InteractiveMarker, InteractiveMarkerControl, InteractiveMarkerFeedback from sensor_msgs.msg import PointCloud2, JointState from ar_track_alvar_msgs.msg import AlvarMarkers from limb_manipulation_msgs.msg import EzgripperAccess, WebAppRequest, WebAppResponse import moveit_commander from moveit_python import PlanningSceneInterface from moveit_msgs.msg import OrientationConstraint import subprocess from robot_controllers_msgs.msg import QueryControllerStatesAction, QueryControllerStatesGoal, ControllerState from database import Database import actionlib import rosbag import os from colour import Color import sys from shared_teleop_functions_and_vars import dpx_to_distance, delta_modified_stamped_pose import copy # maximum times to retry if a transform lookup fails TRANSFROM_LOOKUP_RETRY = 10 # colors for trajectory visualization START_COLOR = Color("Orange") END_COLOR = Color("Blue") TRAJ_HIGHLIGHT_SCALE = Vector3(0.05, 0.008, 0.008) WAYPOINT_HIGHLIGHT_SCALE = Vector3(0.05, 0.01, 0.01) # Vector3(0.055, 0.009, 0.009) WAYPOINT_HIGHLIGHT_COLOR = ColorRGBA(1.0, 0.0, 0.0, 0.8) # body parts and their corresponding ID# and actions BODY_PARTS = {0: "right wrist", 1: "lower right leg", 2: "left wrist", 3: "lower left leg"} ACTIONS = {0: ["right arm elbow extension", "right arm elbow flexion"], 1: ["right leg abduction and adduction"], # , "right leg medial rotation", "right leg lateral rotation"], 2: ["left arm elbow extension", "left arm elbow flexion"], 3: ["left leg abduction and adduction"] #, "left leg medial rotation", "left leg lateral rotation"] } ABBR = {"right leg abduction and adduction": "RLAA", # "right leg medial rotation": "RLMR", "right leg lateral rotation": "RLLR", "left abduction and adduction": "LLAA", # "left leg medial rotation": "LLMR", "left leg lateral rotation": "LLLR", "right arm elbow extension": "RAEE", "right arm elbow flexion": "RAEF", "left arm elbow extension": "LAEE", "left arm elbow flexion": "LAEF" } # leg medial rotation, leg lateral rotation # shoulder flexion, shoulder abduction, shoulder adduction, shoulder medial rotation, shoulder lateral rotation, # forearm pronation, forearm supination, # knee flexion, knee extension (seems to be performed when seated?) def wait_for_time(): """ Wait for simulated time to begin. """ while rospy.Time().now().to_sec() == 0: pass class ArTagReader(object): def __init__(self): self.markers = [] # list of markers (update in real time) self.saved_markers = [] # list of markers saved (update only if update() is called) def callback(self, msg): self.markers = msg.markers def update(self): self.saved_markers = self.markers def get_tag(self, tag): """ Returns the marker with id# == tag """ for marker in self.saved_markers: if marker.id == int(tag): result = PoseStamped() result.pose = marker.pose result.header = marker.header return result return None def get_list(self): """ Returns the list of saved markers """ return self.saved_markers class PbdServer(): """ Server for PBD """ def __init__(self): # controls of Fetch self._arm = fetch_api.Arm() self._arm_joints = fetch_api.ArmJoints() self._torso = fetch_api.Torso() self._head = fetch_api.Head() self._base = fetch_api.Base() self._fetch_gripper = fetch_api.Gripper() # status of the arm: relax or freeze self._arm_relaxed = False # transformation self._tf_listener = tf.TransformListener() rospy.sleep(0.1) # AR tag reader self._reader = ArTagReader() self._ar_sub = rospy.Subscriber("ar_pose_marker", AlvarMarkers, callback=self._reader.callback) # database of actions self._db = Database() self._db.load() # publisher and subscriber for controls of SAKE gripper self._sake_gripper_pub = rospy.Publisher('/ezgripper_access', EzgripperAccess, queue_size=1) self._sake_gripper_sub = rospy.Subscriber('/ezgripper_access_status', EzgripperAccess, callback=self._set_sake_gripper_action_status) # moveit: query controller self._controller_client = actionlib.SimpleActionClient('/query_controller_states', QueryControllerStatesAction) # moveit: move group commander moveit_commander.roscpp_initialize(sys.argv) moveit_robot = moveit_commander.RobotCommander() self._moveit_group = moveit_commander.MoveGroupCommander('arm') # motion planning scene self._planning_scene = PlanningSceneInterface('base_link') self._planning_scene.clear() # visualization self._viz_pub = rospy.Publisher('visualization_marker', Marker, queue_size=5) self._viz_markers_pub = rospy.Publisher('visualization_marker_array', MarkerArray, queue_size=5) # initial position of robot arm self._arm_initial_poses = [ ("shoulder_pan_joint", 1.296), ("shoulder_lift_joint", 1.480), ("upperarm_roll_joint", -0.904), ("elbow_flex_joint", 2.251), ("forearm_roll_joint", -2.021), ("wrist_flex_joint", -1.113), ("wrist_roll_joint", -0.864)] # orientation constraint (unused) self._gripper_oc = OrientationConstraint() self._gripper_oc.header.frame_id = 'base_link' self._gripper_oc.link_name = 'wrist_roll_link' self._gripper_oc.orientation.z = -0.707 self._gripper_oc.orientation.w = 0.707 self._gripper_oc.absolute_x_axis_tolerance = 0.1 self._gripper_oc.absolute_y_axis_tolerance = 0.1 self._gripper_oc.absolute_z_axis_tolerance = 3.14 self._gripper_oc.weight = 1.0 # moveit args self._kwargs = { 'allowed_planning_time': 30, 'execution_timeout': 15, 'group_name': 'arm', 'num_planning_attempts': 10, # 'orientation_constraint': self._gripper_oc, 'replan': True, 'replan_attempts': 5, 'tolerance': 0.01 } # current pose of gripper (used in performing actions) self._current_pose = None # bag file directory script_path = os.path.abspath(__file__) self._bag_file_dir = os.path.split(script_path)[0][:-4] + '/bags' # subscriber and publisher for frontend self._web_app_request_sub = rospy.Subscriber("web_app_request", WebAppRequest, callback=self.web_app_request_callback) self._web_app_response_pub = rospy.Publisher('web_app_response', WebAppResponse, queue_size=5) # variables representing the program state self._sake_gripper_attached = False self._sake_gripper_action_finished = False self._sake_gripper_effort = "100" self._robot_stopped = False self._grasp_position_ready = False self._grasp_type = "h_close" self._do_position_ready = False self._do_position_id = -1 self._preview_action_abbr = "" self._preview_traj = [] # the trajectory being previewed currently self._current_waypoint_id = -1 rospy.sleep(0.5) def setup(self): """ Handler for robot set up """ print("\nSetting up everything, please wait...\n") # set robot's initial state self._torso.set_height(0) self._head.pan_tilt(0, 0.8) # move arm to the initial position (with collision detection) self._arm.move_to_joint_goal(self._arm_initial_poses, replan=True) print("\nThe program is ready to use :-)\n") def shutdown(self): """ Handler for robot shutdown """ print("\nShutting down... Bye :-)\n") # clear display self._viz_markers_pub.publish(MarkerArray(markers=[])) # moveit self.freeze_arm() self._planning_scene.clear() self._arm.cancel_all_goals() # save the database self._db.save() # moveit: move group commander self._moveit_group.stop() moveit_commander.roscpp_shutdown() def attach_sake_gripper(self): """ Attaches SAKE gripper to Fetch's gripper, and updates the planning scene. """ self.freeze_arm() # attach SAKE gripper to Fetch's gripper self._fetch_gripper.open() # make sure Fetch's gripper is open self._fetch_gripper.close() self._sake_gripper_attached = True # add SAKE gripper to the planning scene frame_attached_to = 'gripper_link' frames_okay_to_collide_with = ['gripper_link', 'l_gripper_finger_link', 'r_gripper_finger_link'] package_path = subprocess.check_output("rospack find ezgripper_driver", shell=True).replace('\n','') if rospy.get_param("use_urdf"): # use real sake gripper mesh # palm sake_palm_pose = Pose(Point(-0.01, 0, 0.05), Quaternion(-0.7, 0, 0.7, 0)) sake_palm_mesh_file = package_path + "/meshes/visual/SAKE_Palm_Dual.stl" self._planning_scene.attachMesh('sake_palm', sake_palm_pose, sake_palm_mesh_file, frame_attached_to, touch_links=frames_okay_to_collide_with) # fingers sake_finger_1_pose = Pose(Point(0, -0.03, -0.055), Quaternion(0.5, -0.5, 0.5, 0.5)) sake_finger_1_mesh_file = package_path + "/meshes/visual/SAKE_Finger.stl" self._planning_scene.attachMesh('sake_finger_1', sake_finger_1_pose, sake_finger_1_mesh_file, frame_attached_to, touch_links=frames_okay_to_collide_with) sake_finger_2_pose = Pose(Point(0, 0.03, -0.055), Quaternion(-0.5, -0.5, -0.5, 0.5)) sake_finger_2_mesh_file = package_path + "/meshes/visual/SAKE_Finger.stl" self._planning_scene.attachMesh('sake_finger_2', sake_finger_2_pose, sake_finger_2_mesh_file, frame_attached_to, touch_links=frames_okay_to_collide_with) else: # use a box to represent the sake gripper self._planning_scene.attachBox('sake', 0.03, 0.09, 0.15, 0, 0, -0.05, frame_attached_to, frames_okay_to_collide_with) self._planning_scene.setColor('sake', 1, 0, 1) self._planning_scene.sendColors() # calibrate SAKE gripper self.do_sake_gripper_action("calibrate") def remove_sake_gripper(self): """ Removes SAKE gripper from Fetch's gripper, and updates the planning scene. """ self.freeze_arm() # remove SAKE gripper from Fetch's gripper self._fetch_gripper.close() # make sure Fetch's gripper is close self._fetch_gripper.open() self._sake_gripper_attached = False # remove SAKE gripper from the planning scene if rospy.get_param("use_urdf"): self._planning_scene.removeAttachedObject('sake_palm') self._planning_scene.removeAttachedObject('sake_finger_1') self._planning_scene.removeAttachedObject('sake_finger_2') else: self._planning_scene.removeAttachedObject('sake') def update_env(self, update_octo=True): """ Updates the list of markers, and scan the surroundings to build an octomap. Returns false if the update fails, true otherwise. """ # update markers self._reader.update() # update octomap if update_octo: # clear previous octomap if not self._clear_octomap(): return False # scan the range: pan -0.75~0.75, tilt 0~0.7 for i in range(6): pan = -0.75 + 0.25 * i self._head.pan_tilt(pan, 0) rospy.sleep(2) self._head.pan_tilt(pan, 0.7) rospy.sleep(2) self._head.pan_tilt(pan, 0) # move the head back to initial position self._head.pan_tilt(0, 0.7) return True def do_sake_gripper_action(self, command): # publish response message if command == "calibrate": self._publish_server_response(msg="Calibrating SAKE gripper, please wait...") elif command == "h_close": self._publish_server_response(msg="Hard closing SAKE gripper, please wait...") elif command == "s_close": self._publish_server_response(msg="Soft closing SAKE gripper, please wait...") elif command == "open": self._publish_server_response(msg="Opening SAKE gripper, please wait...") else: # [percentage open, effort] self._publish_server_response(msg="Closing SAKE gripper, please wait...") # publish command self._sake_gripper_pub.publish(EzgripperAccess(type=command)) # wait for the action to finish if in real if not rospy.get_param("use_sim"): while not self._sake_gripper_action_finished: continue # finished, reset self._sake_gripper_action_finished = False def reset(self): """ Moves arm to its initial position and calibrates gripper """ self.freeze_arm() self._arm.move_to_joint_goal(self._arm_initial_poses, replan=True) self.do_sake_gripper_action("calibrate") self._robot_stopped = False self._reset_program_state() def estop(self): """ Emergency stop. """ self.relax_arm() self._robot_stopped = True self._reset_program_state() def preview_body_part_with_id(self, id_num): """ Publishes visualization markers to mark the body part with given id. """ raw_pose = self._get_tag_with_id(id_num) if raw_pose is not None: # visualize goal pose marker = Marker(type=Marker.CUBE, id=0, pose=raw_pose.pose.pose, scale=Vector3(0.06, 0.06, 0.06), header=raw_pose.header, color=ColorRGBA(1.0, 0.75, 0.3, 0.8)) self._viz_pub.publish(marker) def goto_part_with_id(self, id_num): """ Moves arm above a body part specified by id_num. Returns true if succeed, false otherwise. """ self.freeze_arm() raw_pose = self._get_tag_with_id(id_num) if raw_pose is not None: # found marker, move towards it self.do_sake_gripper_action("40 " + self._sake_gripper_effort) # OPTION 1: pregrasp ---> grasp # highlight and move to the pre-grasp pose pre_grasp_offset = self._db.get("PREGRASP") pre_grasp_pose = self._move_arm_relative(raw_pose.pose.pose, raw_pose.header, offset=pre_grasp_offset, preview_only=True) self.highlight_waypoint(pre_grasp_pose, WAYPOINT_HIGHLIGHT_COLOR) if self._move_arm(pre_grasp_pose, final_state=False): # highlight and move to the grasp pose, clear octomap to ignore collision only at this point if self._clear_octomap(): return self._move_arm(self._get_goto_pose(raw_pose), final_state=False, seed_state=self._get_seed_state()) # # OPTION 2: grasp # return self._move_arm(self._get_goto_pose(raw_pose), final_state=False) # marker with id_num is not found, or some error occured return False def preview_action_with_abbr(self, abbr, id_num): """ Publishes visualization markers to preview waypoints on the trajectory with given abbr. Saves waypoints extracted from bag file to database if the entry "abbr" doesn't exist. Returns the colors of waypoints with respect to the ar tag. Records the positions of waypoints with respect to the ar tag. """ # clear previous markers self._viz_markers_pub.publish(MarkerArray(markers=[])) # check the database for the action waypoints = self._db.get(abbr) if waypoints == None or len(waypoints) == 0: waypoints = self._save_traj_to_db(abbr, id_num) self._preview_action_abbr = "" self._preview_traj = [] waypoints_with_respect_to_tag = [] if waypoints: raw_pose = self._get_tag_with_id(id_num) if raw_pose is not None: prev_pose = self._get_goto_pose(raw_pose) # markers marker_arr = [] # marker color gradient colors = list(START_COLOR.range_to(END_COLOR, len(waypoints))) # visualize the trajectory for i in range(len(waypoints)): # visualize the current waypoint marker = Marker(type=Marker.ARROW, id=i, pose=prev_pose.pose, scale=TRAJ_HIGHLIGHT_SCALE, header=prev_pose.header, color=ColorRGBA(colors[i].red, colors[i].green, colors[i].blue, 0.8)) marker_arr.append(marker) # record the waypoint waypoints_with_respect_to_tag.append(str(colors[i].hex)) self._preview_traj.append(prev_pose) if i < len(waypoints) - 1: # calculate offset between the current point on the trajectory and the next point r_pos = waypoints[i].pose # current point r_mat = self._pose_to_transform(r_pos) w_mat = self._pose_to_transform(waypoints[i + 1].pose) offset = np.dot(np.linalg.inv(r_mat), w_mat) prev_pose = self._move_arm_relative(prev_pose.pose, prev_pose.header, offset=offset, preview_only=True) # publish new markers self._viz_markers_pub.publish(MarkerArray(markers=marker_arr)) # record the action name self._preview_action_abbr = abbr return waypoints_with_respect_to_tag def highlight_waypoint(self, highlight_pose, color): """ Publishes a marker at the specified location. """ marker = Marker(type=Marker.ARROW, id=0, pose=highlight_pose.pose, scale=WAYPOINT_HIGHLIGHT_SCALE, header=highlight_pose.header, color=color) self._viz_pub.publish(marker) def edit_waypoint(self, waypoint_id, delta_x, delta_y, camera): """ Temporarily saves the changes to the specified waypoint, and highlights the resulting pose. """ # calculate the resulting pose new_pose = self._compute_pose_by_delta(self._preview_traj[waypoint_id], delta_x, delta_y, camera) # save the new pose self._preview_traj[waypoint_id] = new_pose # preview the new trajectory marker_arr = [] # marker color gradient colors = list(START_COLOR.range_to(END_COLOR, len(self._preview_traj))) # visualize the trajectory for i in range(len(self._preview_traj)): # highlight the waypoint that is being editing color = WAYPOINT_HIGHLIGHT_COLOR if i == waypoint_id else ColorRGBA(colors[i].red, colors[i].green, colors[i].blue, 0.8) marker = Marker(type=Marker.ARROW, id=i, pose=self._preview_traj[i].pose, scale=TRAJ_HIGHLIGHT_SCALE, header=self._preview_traj[i].header, color=color) marker_arr.append(marker) # clear previous markers self._viz_markers_pub.publish(MarkerArray(markers=[])) # publish new markers self._viz_markers_pub.publish(MarkerArray(markers=marker_arr)) def modify_traj_in_db(self, cancel_change=True): """ Overwrites the previous trajectory in database. """ if not cancel_change: self._db.delete(self._preview_action_abbr) self._db.add(self._preview_action_abbr, self._preview_traj) self._preview_action_abbr = "" self._preview_traj = [] def do_action_with_abbr(self, abbr, id_num): """ Moves arm to perform the action specified by abbr, save the trajectory to database if neccessary. Returns true if succeed, false otherwise. """ action_result = False if self._prepare_action(abbr, id_num): action_result = self._follow_traj_step_by_step(0) return action_result def do_action_with_abbr_smooth(self, abbr, id_num): """ Moves arm to perform the action specified by abbr smoothly. Returns true if succeed, false otherwise. """ action_result = False if self._prepare_action(abbr, id_num): # calculate a smooth trajectory passing through all the waypoints and move the arm action_result = self._move_arm(None, trajectory_waypoint=self._preview_traj, final_state=True, seed_state=self._get_seed_state()) if not action_result: # smooth action fails, do the action step by step instead action_result = self._follow_traj_step_by_step(0) return action_result def pause_action(self): """ Pause the current action. """ # self.relax_arm() print(self._current_waypoint_id) def continue_action(self): """ Continue the current action. """ # self.freeze_arm() print(self._current_waypoint_id) # if self._current_waypoint_id > -1: # # continue going to the next waypoint # self._follow_traj_step_by_step(self._current_waypoint_id + 1) def record_action_with_abbr(self, abbr, id_num): """ Records the pose offset named abbr relative to tag, always overwrites the previous entry (if exists). Returns true if succeeds, false otherwise. """ # get tag pose tag_pose = self._get_tag_with_id(id_num) if tag_pose == None: return False # get the pose to be recorded: transformation lookup (position, quaternion) = self._tf_lookup() if (position, quaternion) == (None, None): return False # get the transformation, record it record_pose = Pose() record_pose.position.x = position[0] record_pose.position.y = position[1] record_pose.position.z = position[2] record_pose.orientation.x = quaternion[0] record_pose.orientation.y = quaternion[1] record_pose.orientation.z = quaternion[2] record_pose.orientation.w = quaternion[3] # get the offset between the tag pose and the pose to be recorded t_pos = tag_pose.pose.pose t_mat = self._pose_to_transform(t_pos) w_mat = self._pose_to_transform(record_pose) offset = np.dot(np.linalg.inv(t_mat), w_mat) # add the offset to database self._db.add(abbr, offset) self._db.save() return True def relax_arm(self): """ Relax the robot arm, if the program is running on the real robot """ if not rospy.get_param("use_sim") and not self._arm_relaxed: goal = QueryControllerStatesGoal() state = ControllerState() state.name = 'arm_controller/follow_joint_trajectory' state.state = ControllerState.STOPPED goal.updates.append(state) self._controller_client.send_goal(goal) self._controller_client.wait_for_result() self._arm_relaxed = True def freeze_arm(self): """ Freeze the robot arm, if the program is running on the real robot """ if not rospy.get_param("use_sim") and self._arm_relaxed: goal = QueryControllerStatesGoal() state = ControllerState() state.name = 'arm_controller/follow_joint_trajectory' state.state = ControllerState.RUNNING goal.updates.append(state) self._controller_client.send_goal(goal) self._controller_client.wait_for_result() self._arm_relaxed = False def get_list(self): """ Returns a list of AR tags recognized by the robot. """ return self._reader.get_list() def get_db_list(self): """ Returns list of entries in the database. """ return self._db.list() def get_db_entry(self, entry): """ Returns values associated with the given entry in the database. """ return self._db.get(entry) def delete_db_entry(self, name): """ Delete the database entry with the given name """ self._db.delete(name) self._db.save() def web_app_request_callback(self, msg): """ Parse the request given by the wee application, and call the corresponding functions. """ request_type, request_args = msg.type, msg.args print("type: " + request_type) print("args: " + str(request_args) + "\n") if request_type == "attach": return_msg = "SAKE gripper has already attached!" if not self._sake_gripper_attached: self._publish_server_response(msg="Attaching SAKE gripper...") self.attach_sake_gripper() return_msg = "SAKE gripper attached" self._publish_server_response(status=True, msg=return_msg) elif request_type == "remove": return_msg = "SAKE gripper has already removed!" if self._sake_gripper_attached: self._publish_server_response(msg="Removing SAKE gripper...") self.remove_sake_gripper() return_msg = "SAKE gripper removed" self._publish_server_response(status=True, msg=return_msg) elif not self._sake_gripper_attached: # need to attach SAKE gripper first self._publish_server_response(status=True, msg="Please attach SAKE gripper first!") else: # SAKE gripper has already attached if request_type == "record": self._publish_server_response(msg="Recording the current scene...") if self.update_env(update_octo=bool(request_args[0])): # get the list of body parts and actions parts = self.get_list() parts_info, actions_info = [], [] if len(parts): for part in parts: if part.id in BODY_PARTS and part.id in ACTIONS: parts_info.append(str(part.id) + ":" + BODY_PARTS[part.id]) for action in ACTIONS[part.id]: actions_info.append(str(part.id) + ":" + action + ":" + ABBR[action]) self._publish_server_response(type="parts", args=parts_info) self._publish_server_response(type="actions", status=True, args=actions_info, msg="Scene recorded") else: self._publish_server_response(status=True, msg="Failed to record the current scene!") elif request_type == "prev_id" and len(request_args) == 1: id_num = int(request_args[0]) # convert from string to int self._publish_server_response(status=True, msg="Previewing " + BODY_PARTS[id_num] + "...") self.preview_body_part_with_id(id_num) elif request_type == "prev" and len(request_args) == 2: abbr, id_num = request_args[0], int(request_args[1]) waypoints_with_respect_to_tag = self.preview_action_with_abbr(abbr, id_num) self._publish_server_response(type=request_type, status=True, args=waypoints_with_respect_to_tag, msg="Previewing the action with respect to body part " + BODY_PARTS[id_num] + "...") elif request_type == "highlight" and len(request_args) == 1: waypoint_id = int(request_args[0]) self.highlight_waypoint(self._preview_traj[waypoint_id], WAYPOINT_HIGHLIGHT_COLOR) self._publish_server_response(status=True) elif request_type == "edit" and len(request_args) == 4: waypoint_id, delta_x, delta_y, camera = int(request_args[0]), int(request_args[1]), int(request_args[2]), request_args[3] self.edit_waypoint(waypoint_id, delta_x, delta_y, camera) self._publish_server_response(status=True) elif request_type == "save_edit" or request_type == "cancel_edit": if request_type == "save_edit": self.modify_traj_in_db(cancel_change=False) else: self.modify_traj_in_db() # cancel all the changes self._publish_server_response(status=True) elif request_type == "reset": self._publish_server_response(msg="Resetting...") self.reset() self._publish_server_response(status=True, msg="Done") elif not self._robot_stopped: # moveit controller is running if request_type == "go" and len(request_args) == 1: self._do_position_ready = False id_num = int(request_args[0]) self._publish_server_response(msg="Moving towards body part " + BODY_PARTS[id_num] + "...") if self.goto_part_with_id(id_num): self._grasp_position_ready = True self._do_position_id = id_num self._publish_server_response(status=True, msg="Done, ready to grasp") else: self._publish_server_response(status=True, msg="Fail to move!") elif request_type == "grasp" and self._grasp_position_ready and len(request_args) == 1: self._publish_server_response(msg="Grasping...") self._grasp_type = "h_close" if request_args[0] == "h" else "s_close" self.do_sake_gripper_action(self._grasp_type) self._grasp_position_ready = False self._do_position_ready = True self._publish_server_response(status=True, msg="Grasped") elif request_type == "relax": self._publish_server_response(msg="Relaxing arm...") self.relax_arm() self._publish_server_response(status=True, msg="Arm relaxed") elif request_type == "freeze": self._publish_server_response(msg="Freezing arm...") self.freeze_arm() self._publish_server_response(status=True, msg="Arm froze") elif (request_type == "do" or request_type == "do_s") and len(request_args) > 0: action_abbr = request_args[0] return_msg = "Action failed!" if self._do_position_ready: # performing mode self._publish_server_response(msg="Performing " + action_abbr + "...") result = False if request_type == "do": # step by step result = self.do_action_with_abbr(action_abbr, self._do_position_id) else: # "do_s": smooth result = self.do_action_with_abbr_smooth(action_abbr, self._do_position_id) if result: return_msg = "Action succeeded" else: return_msg = "Unknown action for body part with ID: " + str(self._do_position_id) # always release gripper self._publish_server_response(msg="Releasing the gripper...") self.do_sake_gripper_action("40 " + self._sake_gripper_effort) self._do_position_ready = False self._publish_server_response(status=True, msg=return_msg) elif request_type == "open": self._publish_server_response(msg="Opening the gripper...") self.do_sake_gripper_action("open") self._do_position_ready = False self._publish_server_response(status=True, msg="Gripper opened") elif request_type == "stop": self._publish_server_response(msg="Stopping the robot...") self.estop() self._publish_server_response(status=True, msg="Robot stopped, please \"RESET\" if you want to continue using it") elif request_type == "run" and len(request_args) == 3: ####################################################################################### ############## todo # start execution from the currect step in the action trajectory # start execution from the very beginning self.web_app_request_callback(WebAppRequest(type="go", args=[request_args[0]])) self.web_app_request_callback(WebAppRequest(type="grasp", args=[request_args[1]])) self.web_app_request_callback(WebAppRequest(type="do_s", args=[request_args[2]])) self._publish_server_response(type=request_type, status=True, msg="DONE") elif request_type == "step": return_msg = "Ready!" waypoint_id = int(request_args[0]) if waypoint_id == -1: # goto tag and grasp self.web_app_request_callback(WebAppRequest(type="go", args=[request_args[1]])) self.web_app_request_callback(WebAppRequest(type="grasp", args=[request_args[2]])) else: # move along trajectory self._grasp_type = "h_close" if request_args[1] == "h" else "s_close" return_msg = "Fail to reach waypoint #" + request_args[0] result = self._goto_waypoint_on_traj_with_id(waypoint_id) if waypoint_id == len(self._preview_traj) - 1: # last point self._publish_server_response(msg="Releasing the gripper...") self.do_sake_gripper_action("40 " + self._sake_gripper_effort) self._do_position_ready = False if result: return_msg = "Reached waypoint #" + request_args[0] self._publish_server_response(type=request_type, status=True, args=[request_args[0]], msg=return_msg) elif request_type == "pause": self._publish_server_response(msg="Pausing...") self.pause_action() self._publish_server_response(type=request_type, status=True, msg="PAUSED") elif request_type == "continue": self._publish_server_response(msg="Continuing...") self.continue_action() self._publish_server_response(type=request_type, status=True, msg="CONTINUED") else: self._publish_server_response(status=True, msg="Invalid command :)") def _reset_program_state(self): """ Resets program state. """ self._grasp_position_ready = False self._do_position_ready = False self._do_position_id = -1 def _clear_octomap(self): """ Clears the octomap. Returns true if succeeds, false otherwise. """ rospy.wait_for_service('clear_octomap') try: clear_octo = rospy.ServiceProxy('clear_octomap', Empty) clear_octo() except rospy.ServiceException, e: rospy.logerr('Fail clear octomap: {}'.format(e)) return False return True def _move_arm_relative(self, ref_pose, ref_header, offset=None, preview_only=False, seed_state=None): """ Calculates the coordinate of the goal by adding the offset to the given reference pose, and moves the arm to the goal. If it's only for previewing, returns the goal pose, else returns the result of the movement. """ goal_pose = PoseStamped() goal_pose.header = ref_header if offset is not None: # current pose is valid, perform action t_mat = self._pose_to_transform(ref_pose) # compute the new coordinate new_trans = np.dot(t_mat, offset) pose = self._transform_to_pose(new_trans) goal_pose.pose = pose else: goal_pose.pose = ref_pose if preview_only: return goal_pose else: # move to the goal position while avoiding unreasonable trajectories! # close SAKE gripper again to ensure the limb is grasped self.do_sake_gripper_action(self._grasp_type) # visualize goal pose self.highlight_waypoint(goal_pose, ColorRGBA(1.0, 1.0, 0.0, 0.8)) return goal_pose if self._move_arm(goal_pose, final_state=True, seed_state=seed_state) else None def _move_arm(self, goal_pose, trajectory_waypoint=[], final_state=False, seed_state=None): """ Moves arm to the specified goal_pose. Returns true if succeed, false otherwise. """ error = None if not final_state: # simply go to the goal_pose error = self._arm.move_to_pose_with_seed(goal_pose, seed_state, [], **self._kwargs) # record the current pose because we still have the "do action" step to do self._current_pose = goal_pose else: # go to goal_pose while avoiding unreasonable trajectories! if trajectory_waypoint: # create an array of waypoints waypoints = [] for tw in trajectory_waypoint: waypoints.append(tw.pose) # using trajectory waypoints to perform a smooth motion plan = self._arm.get_cartesian_path(self._moveit_group, seed_state, waypoints) if plan: error = self._arm.execute_trajectory(self._moveit_group, plan) else: error = 'PLANNING_FAILED' else: # using seed error = self._arm.move_to_pose_with_seed(goal_pose, seed_state, [], **self._kwargs) if error is not None: # planning with seed failed, try without seed # moveit: move group commander # check if the pose can be reached in a straight line motion plan = self._arm.straight_move_to_pose_check(self._moveit_group, goal_pose) if plan: error = self._arm.straight_move_to_pose(self._moveit_group, plan) else: error = 'PLANNING_FAILED' # reset current pose to none self._current_pose = None if error is not None: self._arm.cancel_all_goals() rospy.logerr("Fail to move: {}".format(error)) return False # succeed return True def _transform_to_pose(self, matrix): """ Matrix to pose """ pose = Pose() trans_vector = tft.translation_from_matrix(matrix) pose.position = Point(trans_vector[0], trans_vector[1], trans_vector[2]) quartern = tft.quaternion_from_matrix(matrix) pose.orientation = Quaternion(quartern[0], quartern[1], quartern[2], quartern[3]) return pose def _pose_to_transform(self, pose): """ Pose to matrix """ q = pose.orientation matrix = tft.quaternion_matrix([q.x, q.y, q.z, q.w]) matrix[0, 3] = pose.position.x matrix[1, 3] = pose.position.y matrix[2, 3] = pose.position.z return matrix def _get_tag_with_id(self, id_num): """ Returns the AR tag with the given id, returns None if id not found """ tag_pose = self._reader.get_tag(id_num) if tag_pose == None: rospy.logerr("AR tag lookup error: Invalid ID# " + str(id_num)) return tag_pose def _tf_lookup(self): """ Lookups the transformation between "base_link" and "wrist_roll_link" (retry up to TRANSFROM_LOOKUP_RETRY times), and returns the result """ (position, quaternion) = (None, None) count = 0 while True: if (position, quaternion) != (None, None): # lookup succeeds return (position, quaternion) elif count >= TRANSFROM_LOOKUP_RETRY: # exceeds maximum retry times rospy.logerr("Fail to lookup transfrom information between 'base_link' and 'wrist_roll_link'") return (None, None) else: # try to lookup transform information try: (position, quaternion) = self._tf_listener.lookupTransform('base_link', 'wrist_roll_link', rospy.Time(0)) except (tf.LookupException, tf.ConnectivityException, tf.ExtrapolationException): count += 1 continue def _get_goto_pose(self, ar_pose): """ Calculates the grasp pose of gripper given the AR tag pose, returns the gripper pose. """ grasp_offset = self._db.get("GRASP") goal_pose = self._move_arm_relative(ar_pose.pose.pose, ar_pose.header, offset=grasp_offset, preview_only=True) self.highlight_waypoint(goal_pose, WAYPOINT_HIGHLIGHT_COLOR) return goal_pose def _prepare_action(self, abbr, id_num): """ Previews the action trajectory, moves robot arm to the starting position of the action and grasps the limb. Returns true if succeeds, false otherwise. """ # preview action self.preview_action_with_abbr(abbr, id_num) # freeze arm for moveit self.freeze_arm() # get AR tag information tag_pose = self._get_tag_with_id(id_num) if tag_pose == None or self._current_pose == None: return False # move arm to the starting position relative to AR tag if not self._preview_traj or not self.goto_part_with_id(id_num): rospy.logerr("Fail to move to the starting position for action: " + abbr) return False return True def _follow_traj_step_by_step(self, start_id): """ Follows the current trajectory step by step. Returns true if at least one waypoint is reached, false otherwise. """ succeed = False for i in range(start_id, len(self._preview_traj) - 1): goal_pose = self._preview_traj[i + 1] # move arm relative to the previous pose (use seed), skip the current waypoint if the current action fails action_result = self._move_arm_relative(goal_pose.pose, goal_pose.header, seed_state=self._get_seed_state()) # record the current waypoint id self._current_waypoint_id = i + 1 # check action result if action_result is not None: succeed = True # the whole action succeeds if at least one pose is reached else: rospy.logerr("Fail to reach waypoint " + str(i + 1)) # action finished, reset the current waypoint self._current_waypoint_id = -1 return succeed def _goto_waypoint_on_traj_with_id(self, waypoint_id): """ Go to the specific waypoint on the trajectory. Returns true if succeeds, false otherwise. """ succeed = False if -1 < waypoint_id < len(self._preview_traj): goal_pose = self._preview_traj[waypoint_id] action_result = self._move_arm_relative(goal_pose.pose, goal_pose.header, seed_state=self._get_seed_state()) if action_result is not None: succeed = True else: rospy.logerr("Fail to reach waypoint " + str(i + 1)) return succeed def _save_traj_to_db(self, abbr, id_num): """ Checks bag file for the trajectory with given abbr, calculate waypoints and save those points to the database. Returns the calculated waypoints if succeed, none otherwise. """ # check bag files for the trajectory bag_file_path = os.path.join(self._bag_file_dir, abbr.lower() + '.bag') bag = rosbag.Bag(bag_file_path) waypoints = [] prev_msg = [] # get the trajectory from bag file for topic, msg, t in bag.read_messages(topics=['/joint_states']): joint_state = list(msg.position[6:13]) if len(joint_state) != 0 and (len(prev_msg) == 0 or np.abs(np.sum(np.subtract(joint_state, prev_msg))) > rospy.get_param("arm_traj_threshold")): prev_msg = joint_state # use forward kinematics to find the wrist position point = self._arm.compute_fk(msg) if point: # add the result position waypoints.append(point[0]) bag.close() if len(waypoints) < 2: # this trajectory is empty because it only contains the starting point rospy.logerr("Empty trajectory for action: " + abbr) return None # add the result position to database self._db.add(abbr, waypoints) self._db.save() return waypoints def _publish_server_response(self, type="", status=False, args=[], msg=""): """ Publishes the server response message, and prints the message in console if needed. """ if rospy.get_param("console_output"): print(msg) self._web_app_response_pub.publish(WebAppResponse(type=type, status=status, args=args, msg=msg)) def _compute_pose_by_delta(self, current_pose, delta_x, delta_y, camera): """ Computes and returns the new pose with respect to base_link after applying delta_x and delta_y to the current pose in the specified camera view. """ x_distance, y_distance = dpx_to_distance(delta_x, delta_y, camera, current_pose, True) return delta_modified_stamped_pose(x_distance, y_distance, camera, current_pose) def _get_seed_state(self): """ Returns the current arm joint state as the seed used in motion planning. """ seed_state = JointState() seed_state.name = self._arm_joints.names() seed_state.position = self._arm_joints.values() return seed_state def _set_sake_gripper_action_status(self, msg): """ This is the callback of sake gripper status. """ self._sake_gripper_action_finished = True
mit
-8,208,984,794,664,362,000
40.827053
150
0.642281
false
mmktomato/zenmai-bts
test/test_zenmai.py
1
15025
"""Unit test""" import unittest import re import io from datetime import datetime, timedelta from flask import request, session from . import ctx from .zenmai_test_utils import create_issue, create_comment, create_attached_file, create_user, \ login, logout, delete_all_issues from web.models.issue import Issue from web.models.user import User class ZenmaiTestCase(unittest.TestCase): """TestCase class""" def _assert_403(self, data): """Helper method of 403 assertion.""" self.assertIn('403', data) self.assertIn('Forbidden', data) def _assert_issue_detail(self, data, subject, body, pub_date, state_name, attached_file_name): """Helper method of issue detail page assertion. Args: data (string): HTTP Response body. subject (string): Regex string of subject. body (string): Regex string of body. pub_date (datetime): pub_date. state_name (string): Regex string of state name. attached_file_name (string): Regex string of attached file name. """ # subject subject_regex = re.compile('<h1>{}</h1>'.format(subject)) self.assertRegex(data, subject_regex) # body body_regex = re.compile(r'<div class="panel-body">.*<p class="zen-comment-body">{}</p>.*</div>'.format(body), re.DOTALL) self.assertRegex(data, body_regex) # pub_date pub_date_regex = re.compile('<div class="panel-heading">.*{0:%Y-%m-%d %H:%M:%S}.*</div>'.format(pub_date), re.DOTALL) self.assertRegex(data, pub_date_regex) # state_name state_name_regex = re.compile('<span class="label.*">{}</span>'.format(state_name)) self.assertRegex(data, state_name_regex) # attached_file_name attached_file_name_regex = re.compile('<div class="panel-footer">.*download: <a href="/download/\d+/">{}</a>.*</div>'.format(attached_file_name), re.DOTALL) self.assertRegex(data, attached_file_name_regex) def test_get_issue_list(self): """Test case of issue list. (HTTP GET)""" issue = create_issue(subject='test subject.test_get_issue_list.') issue.add() res = ctx['TEST_APP'].get('/') data = res.data.decode('utf-8') subject_regex = re.compile(r'<a href="/{}/">.*test subject\.test_get_issue_list\..*</a>'.format(issue.id)) self.assertRegex(data, subject_regex) def test_get_empty_issue_list(self): """Test case of no issues. (HTTP GET)""" delete_all_issues() res = ctx['TEST_APP'].get('/') data = res.data.decode('utf-8') self.assertIn('No issues.', data) def test_get_issue_detail(self): """Test case of issue detail. (HTTP GET)""" pub_date = datetime.utcnow() + timedelta(days=1) # tommorow attached_file = create_attached_file(name='test.txt') comment = create_comment(body='test body.test_get_issue_detail.', pub_date=pub_date, attached_files=[attached_file]) issue = create_issue(subject='test subject.test_get_issue_detail.', comments=[comment], state_id=2) issue.add() res = ctx['TEST_APP'].get('/{}/'.format(issue.id)) data = res.data.decode('utf-8') self._assert_issue_detail( data=data, subject='test subject\.test_get_issue_detail\.', body='test body\.test_get_issue_detail\.', pub_date=pub_date, state_name=issue.state.name, attached_file_name='test\.txt') def test_post_issue_detail(self): """Test case of issue detail. (HTTP POST)""" issue = create_issue() issue.add() # without authentication. res = ctx['TEST_APP'].post('/{}/'.format(issue.id), data={ 'csrf_token': ctx['CSRF_TOKEN'] }, follow_redirects=True) self._assert_403(res.data.decode('utf-8')) # with authentication. with login() as (user, _): res = ctx['TEST_APP'].post('/{}/'.format(issue.id), data={ 'csrf_token': ctx['CSRF_TOKEN'], 'new_body': 'test body.test_post_issue_detail', 'new_state': 1, 'file': (io.BytesIO(b'test attached file content.test_post_issue_detail.'), 'test.txt') }, follow_redirects=True) data = res.data.decode('utf-8') self._assert_issue_detail( data=data, subject=issue.subject, body=issue.comments[0].body, pub_date=issue.comments[0].pub_date, state_name=issue.state.name, attached_file_name='test\.txt') def test_get_no_issue_detail(self): """Test case of no issue detail. (HTTP GET)""" issue = create_issue() issue.add() res = ctx['TEST_APP'].get('/{}/'.format(issue.id + 1)) self.assertEqual(res.status_code, 404) def test_get_new_issue(self): """Test case of new issue page. (HTTP GET)""" # without authentication. res = ctx['TEST_APP'].get('/new/', follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - login</title>', data) self.assertIn('you need to login.', data) # with authentication. with login() as (_, _): res = ctx['TEST_APP'].get('/new/') data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - new issue</title>', data) self.assertIn('Add new issue', data) def test_post_new_issue(self): """Test case of new issue page. (HTTP POST)""" # without authentication. res = ctx['TEST_APP'].post('/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'] }, follow_redirects=True) self._assert_403(res.data.decode('utf-8')) # with authentication. with login() as (user, _): res = ctx['TEST_APP'].post('/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'new_subject': 'test subject.test_post_new_issue.', 'new_body': 'test body.test_post_new_issue.', 'new_state': 1, 'file': (io.BytesIO(b'test attached file content.test_post_new_issue.'), 'test.txt') }, follow_redirects=True) data = res.data.decode('utf-8') issue = Issue.query \ .filter_by(subject='test subject.test_post_new_issue.') \ .first() self._assert_issue_detail( data=data, subject='test subject\.test_post_new_issue\.', body='test body\.test_post_new_issue\.', pub_date=issue.comments[0].pub_date, state_name='Open', attached_file_name='test\.txt') def test_post_large_attached_file(self): """Test case of post request with too large attached file.""" large_buf = bytes(ctx['APP'].config['MAX_CONTENT_LENGTH'] + 1) res = ctx['TEST_APP'].post('/new/', data={ 'new_subject': 'test subject.test_post_new_issue.', 'new_body': 'test body.test_post_new_issue.', 'new_state': 1, 'file': (io.BytesIO(large_buf), 'test.txt') }, follow_redirects=True) self.assertEqual(res.status_code, 413) def test_get_download_attached_file(self): """Test case of downloading attached file. (HTTP GET)""" attached_file = create_attached_file(data=b'test content of attached file.test_get_download_attached_file.') comment = create_comment(attached_files=[attached_file]) issue = create_issue(comments=[comment]) issue.add() res = ctx['TEST_APP'].get('/download/{}/'.format(attached_file.id)) data = res.data.decode('utf-8') self.assertEqual(data, 'test content of attached file.test_get_download_attached_file.') def test_get_login_page(self): """Test case of login page. (HTTP GET)""" res = ctx['TEST_APP'].get('/user/login/') data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - login</title>', data) self.assertEqual(res.status_code, 200) def test_post_login_page(self): """Test case of login. (HTTP POST)""" user = create_user( \ id='testid.test_post_login_page', \ name='testname.test_post_login_page', \ password='testpassword.test_post_login_page') with login(user, 'testpassword.test_post_login_page') as (_, res): data = res.data.decode('utf-8') self.assertEqual(res.status_code, 200) self.assertIn('<li><a href="/user/">{}(id:{})</a></li>'.format(user.name, user.id), data) def test_get_logout_page(self): """Test case of logout. (HTTP GET)""" user = create_user( \ id='testid.test_get_logout_page', \ name='testname.test_post_logout_page', \ password='testpassword.test_post_logout_page') with login(user, 'testpassword.test_post_logout_page', do_logout=False): pass res = logout() data = res.data.decode('utf-8') self.assertEqual(res.status_code, 200) self.assertIn('<title>Zenmai - issues</title>', data) def test_get_user_register_page(self): """Test case of user register page. (HTTP GET)""" res = ctx['TEST_APP'].get('/user/new/') data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - register</title>', data) self.assertEqual(res.status_code, 200) def test_post_register_valid_user(self): """Test case of valid user registration. (HTTP POST)""" res = ctx['TEST_APP'].post('/user/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': 'testid.test_post_register_valid_user', 'user_name': 'testname.test_post_register_valid_user', 'password': 'testpassword.test_post_register_valid_user', 'password_retype': 'testpassword.test_post_register_valid_user' }, follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - login</title>', data) self.assertEqual(res.status_code, 200) def test_post_register_invalid_user(self): """Test case of invalid user registration. (HTTP POST)""" # password is not matched. res = ctx['TEST_APP'].post('/user/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': 'testid.test_post_register_invalid_user', 'user_name': 'testname.test_post_register_invalid_user', 'password': 'testpassword.test_post_register_invalid_user', 'password_retype': 'invalid password' }, follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - register</title>', data) self.assertIn('password is not matched.', data) self.assertEqual(res.status_code, 200) # already exist. ctx['TEST_APP'].post('/user/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': 'testid.test_post_register_invalid_user', 'user_name': 'testname.test_post_register_invalid_user', 'password': 'testpassword.test_post_register_invalid_user', 'password_retype': 'testpassword.test_post_register_invalid_user' }, follow_redirects=True) res = ctx['TEST_APP'].post('/user/new/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': 'testid.test_post_register_invalid_user', 'user_name': 'testname.test_post_register_invalid_user', 'password': 'testpassword.test_post_register_invalid_user', 'password_retype': 'testpassword.test_post_register_invalid_user' }, follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - register</title>', data) self.assertIn("id &#39;{}&#39; is already exists.".format('testid.test_post_register_invalid_user'), data) self.assertEqual(res.status_code, 200) def test_get_user_page(self): """Test case of user page. (HTTP GET)""" # without authentication. res = ctx['TEST_APP'].get('/user/', follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - login</title>', data) self.assertEqual(res.status_code, 200) # with authentication. with login() as (user, _): res = ctx['TEST_APP'].get('/user/') data = res.data.decode('utf-8') self.assertIn('<td>{}</td>'.format(user.id), data) self.assertIn('<td>{}</td>'.format(user.name), data) self.assertEqual(res.status_code, 200) def test_get_user_edit_page(self): """Test case of user edit page. (HTTP GET)""" # without authentication. res = ctx['TEST_APP'].get('/user/edit/') data = res.data.decode('utf-8') self._assert_403(res.data.decode('utf-8')) # with authentication. with login() as (user, _): res = ctx['TEST_APP'].get('/user/edit/') data = res.data.decode('utf-8') self.assertIn('<title>Zenmai - edit user</title>', data) self.assertIn('value="{}" readonly'.format(user.id), data) self.assertIn('value="{}"'.format(user.name), data) self.assertIn("leave empty if you don't want to change password.", data) self.assertEqual(res.status_code, 200) def test_post_user_edit_page(self): """Test case of edit user. (HTTP POST)""" # without authentication. res = ctx['TEST_APP'].post('/user/edit/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': 'testid.test_post_user_edit_page', 'user_name': 'testname.test_post_user_edit_page', 'password': 'testpassword.test_post_user_edit_page', 'password_retype': 'testpassword.test_post_user_edit_page' }, follow_redirects=True) self._assert_403(res.data.decode('utf-8')) # with authentication. with login() as (user, _): res = ctx['TEST_APP'].post('/user/edit/', data={ 'csrf_token': ctx['CSRF_TOKEN'], 'user_id': user.id, 'user_name': 'new testname.test_post_user_edit_page', 'password': 'new testpassword.test_post_user_edit_page', 'password_retype': 'new testpassword.test_post_user_edit_page' }, follow_redirects=True) data = res.data.decode('utf-8') self.assertIn('<td>{}</td>'.format(user.id), data) self.assertIn('<td>{}</td>'.format('new testname.test_post_user_edit_page'), data) self.assertEqual(res.status_code, 200)
mit
1,231,649,006,724,835,000
40.969274
164
0.568053
false
MrBhendel/Cirrus
ModulusList.py
1
1374
import sys from abc import ABCMeta, abstractmethod class ModulusList: ''' Maintains a list of (host, modulus, e) tuples. ''' __metaclass__ = ABCMeta def __init__(self): self._modulusList = [] def addModulusList(self, other): for i in range(0, other.length()): item = other[i] self.add(item[0], item[1], item[2]) @abstractmethod def add(self, host, modulus, e): pass @abstractmethod def length(self): pass @abstractmethod def __getitem__(self, index): pass @abstractmethod def saveListToFile(self, fileName): pass @abstractmethod def loadListFromFile(self, fileName): pass class ModulusListImpl(ModulusList): def add(self, host, modulus, e): self._modulusList.append((host, modulus, e)) def length(self): return len(self._modulusList) def __getitem__(self, index): return self._modulusList[index] def saveListToFile(self, fileName): saveFile = open(fileName, 'w') for record in self._modulusList: saveFile.write(str(record[0]) + '\n') saveFile.write(str(record[1]) + '\n') saveFile.write(str(record[2]) + '\n') saveFile.close() def loadListFromFile(self, fileName): loadFile = open(fileName, 'r') while True: host = loadFile.readline().rstrip() n = loadFile.readline().rstrip() e = loadFile.readline().rstrip() if not e: break self._modulusList.append((host, long(n), long(e)))
mit
1,942,632,575,379,380,700
19.818182
53
0.6754
false
kshimasaki/ML-fMRI-Pipeline
Python_Code/dendrogram_specific_file.py
1
6075
import scipy import scipy.cluster.hierarchy as sch import matplotlib.pylab as plt import matplotlib.axes import scipy.io as scio import os import numpy import sklearn.metrics.pairwise import scipy.spatial.distance import sys def cluster(network, subjects_or_features, gt_append , method_name): os.chdir('/Volumes/TITAN/Workingdir/Studies/Space_Fortress/Inputs/Neurodata/Features/Learn_Features/Pre/'+network+'_from_'+network+'/') root = '/Volumes/TITAN/Workingdir/Studies/Space_Fortress/Inputs/Neurodata/Features/Learn_Features/Pre/'+network+'_from_'+network+'/' #for root, dirs, files in os.walk('/Volumes/TITAN/Workingdir/Studies/Space_Fortress/Inputs/Neurodata/Features/Learn_Features/Sub/'+network+'_from_'+network+'/Uncontrolled/No_GT/'+gt_append, topdown=False): #for matrix in files: #if matrix.endswith("_Features.mat"): matrix = "All_Features.mat" current = os.path.join(root, matrix) data = scipy.io.loadmat(current) data = data['learnPATS'] #features if subjects_or_features == 0: data = numpy.rot90(data) #subjects n = len(data) p = len(data[0]) sqeuclid = numpy.zeros((n, n)) for i in range(0, n): for j in range(0, n): for k in range(0, p): sqeuclid[i][j] = 0 sqeuclid[i][j] = sqeuclid[i][j] + (data[i][k] - data[j][k])**2 #d = sch.distance.pdist(data, 'euclidean') d = sch.distance.pdist(data) #numpy.savetxt("sqeuclid.csv", sqeuclid, delimiter=",") #proximity_features = scipy.spatial.distance.squareform(sqeuclid) Z= sch.linkage(sqeuclid,method='single') P= sch.dendrogram(Z,truncate_mode='none') fig = plt.gcf() # fig.set_size_inches(125,60) if subjects_or_features == 1: matplotlib.rcParams.update({'font.size': 16}) fig.set_size_inches(40,10) elif subjects_or_features == 0: matplotlib.rcParams.update({'font.size': 10}) fig.set_size_inches(14,9) fig.set_size_inches(9000,100) fig.autofmt_xdate() plt.xticks(rotation=90) temp = '/Volumes/TITAN/Workingdir/Studies/Space_Fortress/Inputs/Neurodata/Features/Learn_Features/Pre/'+network+'_from_'+network+'/' #temp = temp.replace("Uncontrolled/No_GT/LOSO_Standard_with_MS/", "") #temp = temp.replace("Uncontrolled/No_GT/LOSO_Standard_No_MS/", "") #print temp name_matrix=matrix.replace('.mat','') fig.savefig(temp+name_matrix+"_"+method_name+"_Dendrogram.pdf") fig.clf() #print temp+name_matrix+"_"+method_name+"_Dendrogram.pdf" #Mlab_linkage= sch.to_mlab_linkage(Z) #scio.savemat(temp+"/"+name_matrix+"linkage"+".mat",{'Mlab_linkage': Mlab_linkage}) #scio.savemat(temp+"/"+name_matrix+"dendro"+".mat",{'P': P}) #R=sch.inconsistent(Z,d=2) #MR=sch.maxRstat(Z,R,3) #MI=sch.maxinconsts(Z,R) #final_clust_maxmono_5= sch.fcluster(Z,t=5,criterion='maxclust_monocrit',monocrit=MI) #scio.savemat(temp+"/"+name_matrix+"_maxmono_5_clusters"+".mat",{'final_clust_maxmono_5': final_clust_maxmono_5}) #final_clust_mono_1= sch.fcluster(Z,t=1,criterion='monocrit',monocrit=MR) #scio.savemat(temp+"/"+name_matrix+"_mono_1_clusters"+".mat",{'final_clust_mono_1': final_clust_mono_1}) #final_clust_dist_1= sch.fcluster(Z,t=1,criterion='distance') #scio.savemat(temp+"/"+name_matrix+"_dist_1_clusters"+".mat",{'final_clust_dist_1': final_clust_dist_1}) increment = 0 while increment <= 1: final_clust_incon_1=sch.fclusterdata(sqeuclid, increment) import csv table = [[]] temp_row = ["Cluster #", "Data id"] table.append(temp_row) #print sch.leaders(Z,final_clust_incon_1) print final_clust_incon_1 for j in range(1, max(final_clust_incon_1)): print "cluster: ", j for i in range(0, len(final_clust_incon_1)): if final_clust_incon_1[i] == j: temp_row = [j, i] table.append(temp_row) #plot_tree(P, range(10)) if not os.path.exists(temp+name_matrix+"_"+method_name+"_clusters"): os.makedirs(temp+name_matrix+"_"+method_name+"_clusters") with open(temp+name_matrix+"_"+method_name+"_clusters/"+str(increment)+".csv", 'w') as f: wrtr = csv.writer(f) wrtr.writerows(table) increment = increment + 0.25 #numpy.savetxt(temp+name_matrix+"_"+method_name+"_test.csv", table, delimiter=",", fmt="%s") scio.savemat(temp+name_matrix+"_"+method_name+"_incon_1_clusters"+".mat",{'final_clust_incon_1': final_clust_incon_1}) #final_clust_max_5=sch.fcluster(Z,t=5) #scio.savemat(temp+"/"+name_matrix+"_max_5_clusters"+".mat",{'final_clust_max_5': final_clust_max_5}) #clustered_data_inc_1=sch.fclusterdata(Z,t=1) #scio.savemat(temp+"/"+name_matrix+"_data_inc_1_clusters"+".mat",{'clustered_data_inc_1': clustered_data_inc_1}) # os.chdir('/Volumes/Users/Aki_Undergrads/Workingdir/Studies/Space_Fortress/Inputs/Neurodata/Partial_Correlation_Matrices/ML_GT_RS_ProcessedSplit/Pre/Black_from_Black/Uncontrolled/') # mat = scipy.io.loadmat('0018001_matrix.mat') # X = scipy.randn(10,2) # X = mat['Matrix_0018001'] # d = sch.distance.pdist(X) # create all graphs at one time def plot_tree( P, pos=None ): icoord = scipy.array( P['icoord'] ) dcoord = scipy.array( P['dcoord'] ) color_list = scipy.array( P['color_list'] ) xmin, xmax = icoord.min(), icoord.max() ymin, ymax = dcoord.min(), dcoord.max() if pos: icoord = icoord[pos] dcoord = dcoord[pos] color_list = color_list[pos] for xs, ys, color in zip(icoord, dcoord, color_list): plt.plot(xs, ys, color) plt.xlim( xmin-10, xmax + 0.1*abs(xmax) ) plt.ylim( ymin, ymax + 0.1*abs(ymax) ) plt.show() if __name__ == '__main__': testVAR = sys.argv[1] testVAR2 = sys.argv[2] testVAR3 = sys.argv[3] method_name = sys.argv[4] # testVAR=raw_input("What network should be used? ") # testVAR2=raw_input("What analysis should be used? ") # testVAR3=raw_input("subjects or features?") while(1): if testVAR3 == "subjects" or testVAR3 == "Subjects": testVAR3 = 0 break elif testVAR3 == "features" or testVAR3 == "Features": testVAR3 = 1 break else: print "Please type either \"subjects\" or \"features\"" # method_name=raw_input("What should I name this? ") cluster(testVAR, testVAR3, testVAR2, method_name)
gpl-2.0
5,425,072,207,551,465,000
33.913793
207
0.677695
false
pf4d/dolfin-adjoint
dolfin_adjoint/adjrhs.py
1
10667
import libadjoint import backend import ufl import ufl.algorithms import adjglobals import adjlinalg import utils def find_previous_variable(var): ''' Returns the previous instance of the given variable. ''' for timestep in range(var.timestep, -1, -1): prev_var = libadjoint.Variable(var.name, timestep, 0) if adjglobals.adjointer.variable_known(prev_var): prev_var.var.iteration = prev_var.iteration_count(adjglobals.adjointer) - 1 return prev_var raise libadjoint.exceptions.LibadjointErrorInvalidInputs, 'No previous variable found' def _extract_function_coeffs(form): for c in ufl.algorithms.extract_coefficients(form): if isinstance(c, backend.Function): yield c class RHS(libadjoint.RHS): '''This class implements the libadjoint.RHS abstract base class for the Dolfin adjoint. It takes in a form, and implements the necessary routines such as calling the right-hand side and taking its derivative.''' def __init__(self, form): self.form=form if isinstance(self.form, ufl.form.Form): self.deps = [adjglobals.adj_variables[coeff] for coeff in _extract_function_coeffs(self.form)] else: self.deps = [] if isinstance(self.form, ufl.form.Form): self.coeffs = [coeff for coeff in _extract_function_coeffs(self.form)] else: self.coeffs = [] def __call__(self, dependencies, values): if isinstance(self.form, ufl.form.Form): dolfin_dependencies=[dep for dep in _extract_function_coeffs(self.form)] dolfin_values=[val.data for val in values] return adjlinalg.Vector(backend.replace(self.form, dict(zip(dolfin_dependencies, dolfin_values)))) else: # RHS is a adjlinalg.Vector. assert isinstance(self.form, adjlinalg.Vector) return self.form def derivative_action(self, dependencies, values, variable, contraction_vector, hermitian): if contraction_vector.data is None: return adjlinalg.Vector(None) if isinstance(self.form, ufl.form.Form): # Find the dolfin Function corresponding to variable. dolfin_variable = values[dependencies.index(variable)].data dolfin_dependencies = [dep for dep in _extract_function_coeffs(self.form)] dolfin_values = [val.data for val in values] current_form = backend.replace(self.form, dict(zip(dolfin_dependencies, dolfin_values))) trial = backend.TrialFunction(dolfin_variable.function_space()) d_rhs = backend.derivative(current_form, dolfin_variable, trial) if hermitian: action = backend.action(backend.adjoint(d_rhs), contraction_vector.data) else: action = backend.action(d_rhs, contraction_vector.data) return adjlinalg.Vector(action) else: # RHS is a adjlinalg.Vector. Its derivative is therefore zero. return adjlinalg.Vector(None) def second_derivative_action(self, dependencies, values, inner_variable, inner_contraction_vector, outer_variable, hermitian, action_vector): if isinstance(self.form, ufl.form.Form): # Find the dolfin Function corresponding to variable. dolfin_inner_variable = values[dependencies.index(inner_variable)].data dolfin_outer_variable = values[dependencies.index(outer_variable)].data dolfin_dependencies = [dep for dep in _extract_function_coeffs(self.form)] dolfin_values = [val.data for val in values] current_form = backend.replace(self.form, dict(zip(dolfin_dependencies, dolfin_values))) trial = backend.TrialFunction(dolfin_outer_variable.function_space()) d_rhs = backend.derivative(current_form, dolfin_inner_variable, inner_contraction_vector.data) d_rhs = ufl.algorithms.expand_derivatives(d_rhs) if len(d_rhs.integrals()) == 0: return None d_rhs = backend.derivative(d_rhs, dolfin_outer_variable, trial) d_rhs = ufl.algorithms.expand_derivatives(d_rhs) if len(d_rhs.integrals()) == 0: return None if hermitian: action = backend.action(backend.adjoint(d_rhs), action_vector.data) else: action = backend.action(d_rhs, action_vector.data) return adjlinalg.Vector(action) else: # RHS is a adjlinalg.Vector. Its derivative is therefore zero. raise exceptions.LibadjointErrorNotImplemented("No derivative method for constant RHS.") def dependencies(self): return self.deps def coefficients(self): return self.coeffs def __str__(self): return hashlib.md5(str(self.form)).hexdigest() class NonlinearRHS(RHS): '''For nonlinear problems, the source term isn't assembled in the usual way. If the nonlinear problem is given as F(u) = 0, we annotate it as M.u = M.u - F(u) . So in order to actually assemble the right-hand side term, we first need to solve F(u) = 0 to find the specific u, and then multiply that by the mass matrix.''' def __init__(self, form, F, u, bcs, mass, solver_parameters, J): '''form is M.u - F(u). F is the nonlinear equation, F(u) := 0.''' RHS.__init__(self, form) self.F = F self.u = u self.bcs = bcs self.mass = mass self.solver_parameters = solver_parameters self.J = J or backend.derivative(F, u) # We want to mark that the RHS term /also/ depends on # the previous value of u, as that's what we need to initialise # the nonlinear solver. var = adjglobals.adj_variables[self.u] self.ic_var = None if backend.parameters["adjoint"]["fussy_replay"]: can_depend = True try: prev_var = find_previous_variable(var) except: can_depend = False if can_depend: self.ic_var = prev_var self.deps += [self.ic_var] self.coeffs += [u] else: self.ic_copy = backend.Function(u) self.ic_var = None def __call__(self, dependencies, values): assert isinstance(self.form, ufl.form.Form) ic = self.u.function_space() # by default, initialise with a blank function in the solution FunctionSpace if hasattr(self, "ic_copy"): ic = self.ic_copy replace_map = {} for i in range(len(self.deps)): if self.deps[i] in dependencies: j = dependencies.index(self.deps[i]) if self.deps[i] == self.ic_var: ic = values[j].data # ahah, we have found an initial condition! else: replace_map[self.coeffs[i]] = values[j].data current_F = backend.replace(self.F, replace_map) current_J = backend.replace(self.J, replace_map) u = backend.Function(ic) current_F = backend.replace(current_F, {self.u: u}) current_J = backend.replace(current_J, {self.u: u}) vec = adjlinalg.Vector(None) vec.nonlinear_form = current_F vec.nonlinear_u = u vec.nonlinear_bcs = self.bcs vec.nonlinear_J = current_J return vec def derivative_action(self, dependencies, values, variable, contraction_vector, hermitian): '''If variable is the variable for the initial condition, we want to ignore it, and set the derivative to zero. Assuming the solver converges, the sensitivity of the solution to the initial condition should be extremely small, and computing it is very difficult (one would have to do a little adjoint solve to compute it). Even I'm not that fussy.''' if variable == self.ic_var: deriv_value = values[dependencies.index(variable)].data return adjlinalg.Vector(None, fn_space=deriv_value.function_space()) else: return RHS.derivative_action(self, dependencies, values, variable, contraction_vector, hermitian) def second_derivative_action(self, dependencies, values, inner_variable, inner_contraction_vector, outer_variable, hermitian, action): '''If variable is the variable for the initial condition, we want to ignore it, and set the derivative to zero. Assuming the solver converges, the sensitivity of the solution to the initial condition should be extremely small, and computing it is very difficult (one would have to do a little adjoint solve to compute it). Even I'm not that fussy.''' if inner_variable == self.ic_var or outer_variable == self.ic_var: deriv_value = values[dependencies.index(outer_variable)].data return adjlinalg.Vector(None, fn_space=deriv_value.function_space()) else: return RHS.second_derivative_action(self, dependencies, values, inner_variable, inner_contraction_vector, outer_variable, hermitian, action) def derivative_assembly(self, dependencies, values, variable, hermitian): replace_map = {} for i in range(len(self.deps)): if self.deps[i] == self.ic_var: continue j = dependencies.index(self.deps[i]) replace_map[self.coeffs[i]] = values[j].data diff_var = values[dependencies.index(variable)].data current_form = backend.replace(self.form, replace_map) deriv = backend.derivative(current_form, diff_var) if hermitian: deriv = backend.adjoint(deriv) bcs = [utils.homogenize(bc) for bc in self.bcs if isinstance(bc, backend.DirichletBC)] + [bc for bc in self.bcs if not isinstance(bc, backend.DirichletBC)] else: bcs = self.bcs return adjlinalg.Matrix(deriv, bcs=bcs) def adj_get_forward_equation(i): (fwd_var, lhs, rhs) = adjglobals.adjointer.get_forward_equation(i) # We needed to cheat the annotation when we registered a nonlinear solve. # However, if we want to actually differentiate the form (e.g. to compute # the dependency of the form on a ScalarParameter) we're going to need # the real F(u) = 0 back again. So let's fetch it here: if hasattr(rhs, 'nonlinear_form'): lhs = rhs.nonlinear_form fwd_var.nonlinear_u = rhs.nonlinear_u rhs = 0 else: lhs = lhs.data rhs = rhs.data return (fwd_var, lhs, rhs)
lgpl-3.0
-4,020,333,220,749,036,500
38.507407
167
0.630355
false
loadimpact/http2-test
fabfile.py
1
11989
""" We need two special hostnames to connect to: StationA and StationB """ from __future__ import with_statement, print_function from fabric.api import local, settings, abort, run, sudo, cd, hosts, env, execute from fabric.contrib.console import confirm from fabric.operations import put, get from fabric.contrib.project import rsync_project import re import subprocess as sp import os.path from StringIO import StringIO current_dir = os.path.dirname(os.path.realpath(__file__)) VNC_LICENSE = [ "xxxxx-xxxxx-xxxxx-xxxxx-xxxxx" ] # TODO: Put a proper deployment mechanism here. env.key_filename = '/home/alcides/.ssh/zunzun_ec2_keypair_0.pem' StationA_H = '[email protected]' StationB_H = '[email protected]' Beefy_H = '[email protected]' # TODO: Make the IP number below deployment-specific... Beefy_InternalIP = '192.168.112.131' StationA_InternalIP = '192.168.112.129' StationB_InternalIP = '192.168.112.130' HomeDir_Name = "ubuntu" @hosts(StationA_H) def StationA(): """ Copies code to StationA """ rsync_project( local_dir = "scripts/StationA", remote_dir = ("/home/{HomeDir_Name}/".format(HomeDir_Name=HomeDir_Name)) ) run("ln -sf /home/{HomeDir_Name}/StationA/onstartup.py /home/{HomeDir_Name}/onstartup.py".format(HomeDir_Name=HomeDir_Name)) @hosts(StationB_H) def setup_dns_masq(): sudo("apt-get install -y dnsmasq") put( local_path = StringIO("addn-hosts=/home/{HomeDir_Name}/dnsmasq_more.conf\n".format(HomeDir_Name=HomeDir_Name)), remote_path = "/etc/dnsmasq.conf", use_sudo=True) @hosts(StationB_H) def StationB(): """ Copies both the chrome plugin and the DNSMasq watcher """ rsync_project( local_dir = "scripts/StationB", remote_dir = ("/home/{HomeDir_Name}/".format(HomeDir_Name=HomeDir_Name)) ) rsync_project( local_dir = "scripts/StationA/chrome_captures_hars", remote_dir = (("/home/{HomeDir_Name}/StationB/".format(HomeDir_Name=HomeDir_Name)).format(HomeDir_Name=HomeDir_Name)) ) run("ln -sf /home/{HomeDir_Name}/StationB/onstartup.py /home/{HomeDir_Name}/onstartup.py".format(HomeDir_Name=HomeDir_Name)) @hosts(StationB_H) def install_updatednsmasq_service(): with settings(warn_only=True): sudo("service updatednsmasq stop") put( local_path = "scripts/StationB/configure_dnsmasq.py", remote_path = "/home/{HomeDir_Name}/StationB/configure_dnsmasq.py".format(HomeDir_Name=HomeDir_Name) , use_sudo = True ) put( local_path = StringIO(""" description "Update dnsmasq" start on runlevel [2345] stop on runlevel [!2345] umask 022 console log env PATH=/opt/openssl-1.0.2/bin/:/usr/bin:/usr/local/bin:/usr/sbin:/bin export PATH env LD_LIBRARY_PATH=/opt/openssl-1.0.2/lib export LD_LIBRARY_PATH env USER={HomeDir_Name} export USER script exec /usr/bin/python /home/{HomeDir_Name}/StationB/configure_dnsmasq.py end script """.format(HomeDir_Name=HomeDir_Name)), remote_path = "/etc/init/updatednsmasq.conf", use_sudo=True ) sudo("service updatednsmasq start") @hosts(Beefy_H) def Beefy(): sudo("apt-get update") sudo("apt-get -y install libgmp-dev") @hosts(Beefy_H) def BeefyRehMimic(): with settings(warn_only=True): sudo("service mimic stop") put( local_path = "dist/build/reh-mimic/reh-mimic", remote_path = "/home/{HomeDir_Name}/reh-mimic".format(HomeDir_Name=HomeDir_Name) ) run("chmod ugo+x /home/{HomeDir_Name}/reh-mimic".format(HomeDir_Name=HomeDir_Name)) sudo("rm /home/{HomeDir_Name}/mimic -rf".format(HomeDir_Name=HomeDir_Name) ) rsync_project( local_dir = "mimic", remote_dir = "/home/{HomeDir_Name}/".format(HomeDir_Name=HomeDir_Name), ) put( local_path = "scripts/mimic.conf", remote_path = "/etc/init/mimic.conf", use_sudo = True ) sudo("touch /root/.rnd") sudo("service mimic start") @hosts(Beefy_H, StationA_H, StationB_H ) def configure_logging(): if env.host_string == StationA_H: put( local_path = StringIO("""$template Logentries,"199fb2e1-8227-4f73-9150-70a34a5d5e0c %HOSTNAME% %syslogtag%%msg%\\n" *.* @@api.logentries.com:10000;Logentries"""), remote_path = "/etc/rsyslog.d/70-logentries.conf", use_sudo = True ) elif env.host_string == StationB_H: put( local_path = StringIO("""$template Logentries,"3d2fd756-407a-4764-b130-1dd6f22a1b62 %HOSTNAME% %syslogtag%%msg%\\n" *.* @@api.logentries.com:10000;Logentries"""), remote_path = "/etc/rsyslog.d/70-logentries.conf", use_sudo = True ) else: put( local_path = StringIO("""$template Logentries,"7551d4e0-fa76-466f-8547-8c9a347a9363 %HOSTNAME% %syslogtag%%msg%\\n" *.* @@api.logentries.com:10000;Logentries"""), remote_path = "/etc/rsyslog.d/70-logentries.conf", use_sudo = True ) sudo("service rsyslog restart") # Check logging works... sudo("logger -t test Hello there Logentries") @hosts (StationA_H, StationB_H) def deploy_specific(): if env.host_string == StationA_H: print("StationA deploy") StationA() elif env.host_string == StationB_H: print("StationB deploy") StationB() else: print("Beefy station deploy") Beefy() @hosts(StationA_H, StationB_H) def apt_stations(): sudo("apt-get update") sudo("apt-get install -y xutils xbase-clients xfonts-base xfonts-75dpi xfonts-100dpi") sudo("apt-get install -y python-pip") sudo("apt-get install -y xdotool") sudo("apt-get install -y xfwm4") @hosts(StationA_H, StationB_H) def pythonlibs(): sudo("pip install python-daemon>=2.0") sudo("pip install raven") @hosts(Beefy_H, StationA_H, StationB_H) def ssl(): """ Copies Openssl and curl to the target hosts... """ sudo("mkdir -p /opt/openssl-1.0.2/") sudo(("chown {HomeDir_Name} /opt/openssl-1.0.2/".format(HomeDir_Name=HomeDir_Name))) rsync_project( local_dir = "/opt/openssl-1.0.2", remote_dir = "/opt/", extra_opts="-avz" ) put( local_path = "scripts/ca-certificates.crt", remote_path = "/etc/ssl/certs/ca-certificates.crt", use_sudo = True ) @hosts(Beefy_H, StationA_H, StationB_H) def ca(): """ Copies the ca certificate to the home... """ put( local_path = "mimic-here/config/ca/cacert.pem", remote_path = ("/home/{HomeDir_Name}/cacert.pem".format(HomeDir_Name=HomeDir_Name)), use_sudo = True ) @hosts(StationA_H, StationB_H) def install_vnc(): """ """ # run("curl -L -o VNC.tar.gz https://www.realvnc.com/download/binary/1720/") # run("tar xvf VNC-5.2.3-Linux-x64-ANY.tar.gz") put( local_path = "scripts/VNC-5.2.3-Linux-x64-ANY.tar.gz", remote_path = ("/home/{HomeDir_Name}/VNC-5.2.3-Linux-x64-ANY.tar.gz".format(HomeDir_Name=HomeDir_Name)), use_sudo = True ) run(("tar -xzf /home/{HomeDir_Name}/VNC-5.2.3-Linux-x64-ANY.tar.gz".format(HomeDir_Name=HomeDir_Name))) # Get a handier name.... run("rm -rf vnc") run(("mv /home/{HomeDir_Name}/VNC-5.2.3-Linux-x64 /home/{HomeDir_Name}/vnc".format(HomeDir_Name=HomeDir_Name))) sudo(("/home/{HomeDir_Name}/vnc/vnclicense -add {VncLicense}".format( HomeDir_Name= HomeDir_Name, VncLicense = VNC_LICENSE[0] ))) # Will demand some for of interactive input... run(("mkdir -p /home/{HomeDir_Name}/.vnc/".format(HomeDir_Name=HomeDir_Name))) run(("mkdir -p /home/{HomeDir_Name}/.vnc/config.d/".format(HomeDir_Name=HomeDir_Name))) sudo(("/home/{HomeDir_Name}/vnc/vncpasswd /home/{HomeDir_Name}/.vnc/config.d/Xvnc".format(HomeDir_Name=HomeDir_Name))) vnc_fix_permissions() @hosts(StationA_H, StationB_H) def vnc_fix_permissions(): sudo(("chown {HomeDir_Name} /home/{HomeDir_Name}/.vnc -R").format(HomeDir_Name=HomeDir_Name)) @hosts(StationA_H, StationB_H) def install_vnc_xstartup(): run(("mkdir -p /home/{HomeDir_Name}/.vnc/".format(HomeDir_Name=HomeDir_Name))) run(("mkdir -p /home/{HomeDir_Name}/.vnc/config.d/".format(HomeDir_Name=HomeDir_Name))) put( local_path = "scripts/vnc-xstartup", remote_path = ("/home/{HomeDir_Name}/.vnc/xstartup".format(HomeDir_Name=HomeDir_Name)) ) run("chmod ugo+x /home/{HomeDir_Name}/.vnc/xstartup".format(HomeDir_Name=HomeDir_Name)) put( local_path = "scripts/xvncfontpath", remote_path = ("/home/{HomeDir_Name}/.vnc/config.d/xvncfontpath".format(HomeDir_Name=HomeDir_Name)) ) @hosts(StationA_H, StationB_H) def setup_google_chrome(): put( local_path = "scripts/google-chrome-stable_current_amd64.deb", remote_path = ("/home/{HomeDir_Name}/google-chrome-stable_current_amd64.deb".format(HomeDir_Name=HomeDir_Name)), use_sudo = True ) really_setup_google_chrome() @hosts(Beefy_H, StationA_H, StationB_H) def ensure_local_hosts(): # Get the contents of /etc/hosts local_file = StringIO() get( local_path = local_file, remote_path = "/etc/hosts", use_sudo = True ) hosts_file = local_file.getvalue() snippet = """# DO NOT EDIT BELOW BY HAND {Beefy_InternalIP} instr.httpdos.com 192.168.112.129 ip-192-168-112-129 192.168.112.130 ip-192-168-112-130 192.168.112.131 ip-192-168-112-131 # END DO NOT EDIT BELOW""".format( StationA_InternalIP = StationA_InternalIP, Beefy_InternalIP = Beefy_InternalIP ) mo = re.search(r"# DO NOT EDIT BELOW BY HAND\n(.*?)\n# END DO NOT EDIT BELOW", hosts_file, re.MULTILINE) if mo: part_before = hosts_file[:mo.start(0)] part_after = hosts_file[mo.end(0):] hosts_file = part_before + snippet + part_after else: hosts_file += "\n" + snippet put( local_path = StringIO(hosts_file), remote_path = "/etc/hosts", use_sudo = True ) @hosts(StationA_H, StationB_H) def really_setup_google_chrome(): sudo("apt-get update") sudo(("apt-get -f install -y".format(HomeDir_Name=HomeDir_Name))) sudo("apt-get install -y --fix-missing xdg-utils") sudo(("dpkg -i --force-depends /home/{HomeDir_Name}/google-chrome-stable_current_amd64.deb".format(HomeDir_Name=HomeDir_Name))) sudo(("apt-get -f install -y".format(HomeDir_Name=HomeDir_Name))) @hosts(StationA_H, StationB_H) def setup_vnc_service(): put( local_path = "scripts/vncserv-{HomeDir_Name}.conf".format(HomeDir_Name=HomeDir_Name), remote_path = "/etc/init/vncserv.conf", use_sudo = True ) put( local_path = "scripts/undaemon.py", remote_path = "/home/{HomeDir_Name}/undaemon.py".format(HomeDir_Name=HomeDir_Name) ) run("chmod ugo+x /home/{HomeDir_Name}/undaemon.py".format(HomeDir_Name=HomeDir_Name)) with settings(warn_only=True): sudo( "service vncserv start" ) @hosts(StationA_H, StationB_H) def disable_lightdm(): contents = StringIO("manual") put( local_path = contents, remote_path = "/etc/init/lightdm.override", use_sudo=True ) @hosts(StationA_H, StationB_H) def touch_xauthority(): run("touch $HOME/.Xauthority") @hosts(StationA_H, StationB_H) def deploy(): execute(apt_stations) execute(setup_dns_masq) execute(setup_google_chrome) execute(deploy_specific) execute(touch_xauthority) execute(disable_lightdm) execute(StationA) execute(StationB) execute(Beefy) execute(ca) execute(ssl) execute(install_vnc) execute(install_vnc_xstartup) execute(ensure_local_hosts) execute(setup_vnc_service) execute(pythonlibs) execute(BeefyRehMimic) execute(install_updatednsmasq_service)
bsd-3-clause
-2,658,566,617,386,937,000
29.820051
131
0.63967
false
iceflow/aws-demo
s3/s3-bucket-copy/existing-bucket-copy/solutions/2-different-account-same-region/check_sqs_list.py
1
1564
#!/usr/bin/python # -*- coding: utf8 -*- from pprint import pprint import sys,os import random import json import gzip import random import boto3 s3 = boto3.resource('s3') client = boto3.client('sqs') QUEUE_ENDPOINT='https://eu-west-1.queue.amazonaws.com/888250974927/s3-copy-list' DST_BUCKET='ireland-leo-test' def check_queue_status(qurl): #print('check_queue_status(%s)'%(qurl)) #return {'number':0} response = client.get_queue_attributes( QueueUrl=qurl, AttributeNames=[ 'All' ] ) #pprint(response) #{u'Attributes': {'ApproximateNumberOfMessages': '1', message_number=0 if 'Attributes' in response: if 'ApproximateNumberOfMessages' in response['Attributes'] and 'ApproximateNumberOfMessagesNotVisible' in response['Attributes']: message_number=int(response['Attributes']['ApproximateNumberOfMessages']) not_visiable_message_number=int(response['Attributes']['ApproximateNumberOfMessagesNotVisible']) if message_number>0 or not_visiable_message_number>0: #print('%04d/%04d : %s'%(message_number, not_visiable_message_number, qurl)) pass return {'number':message_number} if __name__ == '__main__': qurl_endpoint=sys.argv[1] q_number=int(sys.argv[2]) total_number=0 for pos in xrange(q_number): response = check_queue_status('{0}-{1}'.format(qurl_endpoint, pos+1)) total_number+=response['number'] print total_number*100 sys.exit(0)
gpl-3.0
2,473,053,559,471,726,000
25.066667
137
0.641944
false
burunduk3/testsys
console.py
1
19930
#!/usr/bin/env python3 import argparse, select, socket, sys, termios, time from dts import config from dts.protocol import Packet, PacketParser class Console: NORMAL, LOCK, WAIT = range(3) def __init__( self, tab, command ): self.state = Console.NORMAL self.value = None self.__line = '' self.__message = '' self.__buffer = '' self.__tab = tab self.__command = command self.__start = 0 self.__cursor = 0 self.__width = 80 # TODO: calculate width using termios and renew it using SIGWINCH self.__history = [] self.__current = [x for x in self.__history] + [self.__line] self.__position = len(self.__history) self.__input = None def __enter( self ): if not self.__line: return self.__command(self.__line) if self.__line in self.__history: self.__history.remove(self.__line) self.__history.append(self.__line) if len(self.__history) > 100: del self.__history[0] self.__line = '' self.__cursor = 0 self.__current = [x for x in self.__history] + [self.__line] self.__position = len(self.__history) def lock( self, message, value ): self.value = value self.state = Console.LOCK self.__message = message print('\r\033[K' + self.__message, end='') def unlock( self, message, wait=True ): self.state = Console.WAIT if wait else Console.NORMAL self.__message = message print('\r\033[K' + self.__message, end='') if not wait: self.write(self.__buffer) self.__buffer = '' self.redraw () def write( self, text, force = False ): if len(text) != 0 and text[-1] != '\n': text += '%\n' # zsh style if self.state is Console.LOCK or force: print('\r\033[K' + text + self.__message, end='') elif self.state is Console.WAIT: self.__buffer += text else: print('\r\033[K' + text, end='') self.redraw() sys.stdout.flush() def redraw( self ): assert self.state is Console.NORMAL if self.__cursor < self.__start: self.__start = self.__cursor if self.__cursor - self.__start + 2 >= self.__width: self.__start = self.__cursor + 3 - self.__width if self.__cursor != len(self.__line) and self.__cursor - self.__start + 3 == self.__width: self.__start += 1 start = '\033[1m<\033[0m' if self.__start != 0 else ' ' finish = '\033[1m>\033[0m' if self.__start + self.__width - 2 < max(len(self.__line), self.__cursor + 1) else '' visual = self.__line[self.__start:self.__start + self.__width - (3 if finish != '' else 2)] move = self.__start + len(visual) + (1 if finish != '' else 0) - self.__cursor print('\r\033[K>' + start + visual + finish + ('\033[%dD' % move if move != 0 else ''), end='') def add( self, text ): if self.state is Console.LOCK: return for key in text: # TODO: setup keys for different terminals if self.__input is not None: self.__input += key if self.__input == "[A": if self.__position != 0: self.__current[self.__position] = self.__line self.__position -= 1 self.__line = self.__current[self.__position] self.__cursor = len(self.__line) self.__start = 0 elif self.__input == "[B": if self.__position != len(self.__history): self.__current[self.__position] = self.__line self.__position += 1 self.__line = self.__current[self.__position] self.__cursor = len(self.__line) self.__start = 0 elif self.__input == "[C": self.__cursor = min(len(self.__line), self.__cursor + 1) elif self.__input == "[D": self.__cursor = max(0, self.__cursor - 1) elif self.__input == "[1;5C": pass # Ctrl+← elif self.__input == "[1;5D": pass # Ctrl+→ elif self.__input == "[1~": self.__cursor = 0 elif self.__input == "[3~": self.__line = self.__line[0:self.__cursor] + self.__line[self.__cursor + 1:] elif self.__input == "[4~": self.__cursor = len(self.__line) elif len(self.__input) > 5: self.write("[debug] unknown escape sequence: \e%s\n" % self.__input) else: continue self.__input = None continue if self.state is Console.WAIT: self.state = Console.NORMAL self.write(self.__buffer) self.__buffer = '' continue if ord(key) >= 0x20 and ord(key) != 0x7f: self.__line = self.__line[0:self.__cursor] + key + self.__line[self.__cursor:] self.__cursor += 1 elif ord(key) == 0x09: bonus, result = self.__tab(self.__line[0:self.__cursor]) if bonus: result += ' ' if result is not None: self.__line = result + self.__line[self.__cursor:] self.__cursor = len(result) elif ord(key) == 0x0a: self.__enter() elif ord(key) == 0x23: pass # Ctrl+W elif ord(key) == 0x7f: if self.__cursor == 0: continue; self.__cursor -= 1 self.__line = self.__line[0:self.__cursor] + self.__line[self.__cursor + 1:] elif ord(key) == 0x1b: self.__input = '' else: global count count += 1 self.write("[debug] count = %d, key=%d\n" % (count, ord(key)), force=True) self.redraw() def tab( line ): # TODO: optimize with prefix tree commands = config.commands targets = sorted([x for x in commands if x.startswith(line)]) if len(targets) == 0: return (False, None) if len(targets) == 1: return (True, targets[0]) index = 0 while index < len(targets[0]) and index < len(targets[1]) and targets[0][index] == targets[1][index]: index += 1 if index > len(line): return (False, targets[0][0:index]) console.write(' '.join(targets) + '\n') return (False, None) def console_command( line ): queue.append((command, line)) def command( line ): global command_id packet_id = "id_%08d" % command_id command_id += 1 console.lock("*** waiting for testsys answer ***", packet_id) s.send(Packet({'ID': packet_id, 'Command': line})()) parser = argparse.ArgumentParser(description="Text console for testsys.") parser.add_argument('--password-file', '-p', action='store', dest='key_file', required=True) parser.add_argument('--name', '-n', action='store', dest='name', help='Displayed name.') parser.add_argument('--contest', '-c', action='store', dest='contest', help='Select contest.') parser.add_argument('--msglevel', '-m', action='store', dest='msglevel', help='Initial testsys verbosity level.') parser.add_argument('server', metavar='<host>:<port>', help='Address of testsys server.') args = parser.parse_args() fd = sys.stdin.fileno() tty_attr_old = termios.tcgetattr(fd) attributes = termios.tcgetattr(fd) attributes[3] &= ~(termios.ECHO | termios.ICANON) attributes[6][termios.VMIN] = 0 termios.tcsetattr(fd, termios.TCSADRAIN, attributes) console = Console(tab, console_command) console.lock('*** connecting to testsys ***', "id_pre0") command_id = 0 reconnect_id = 0 with open(args.key_file, 'rb') as key_file: key = key_file.read().decode(config.encoding) assert len(key) == 256, "invalid key: should be 256 bytes length" # TODO: default port 17240, or not host, port = args.server.split(':') port = int(port) poll = select.epoll() poll.register(sys.stdin, select.EPOLLIN) try: s = socket.socket() poll.register(s, select.EPOLLIN) s.connect((host, port)) s.send(Packet({'Password': key, 'Command': "ver", 'ID': "id_pre0"})()) if args.msglevel is not None: console.value = "id_pre1" s.send(Packet({'Command': "msg_level " + args.msglevel, 'ID': "id_pre1"})()) if args.name is not None: console.value = "id_pre2" s.send(Packet({'Command': "name " + args.name, 'ID': "id_pre2"})()) if args.contest is not None: console.value = "id_pre3" s.send (Packet ({'Command': "select_contest " + args.contest, 'ID': "id_pre3"})()) except KeyboardInterrupt: console.lock("terminated by KeyboardInterrupt", "never") print(""); termios.tcsetattr(fd, termios.TCSADRAIN, tty_attr_old) sys.exit(1) def reconnect(): global reconnect_id, is_reconnect, s is_reconnect = True packet_id = "reconnect%d" % reconnect_id console.lock("*** reconnecting ***", packet_id + "_2") reconnect_id += 1 del action[s.fileno()] try: poll.unregister(s) s.close() time.sleep(1) s = socket.socket() poll.register(s, select.EPOLLIN) action[s.fileno()] = handle_socket console.write("", force=True) s.connect((host, port)) s.send(Packet({'Password': key, 'Command': "ver", 'ID': packet_id + "_0"})()) if args.msglevel is not None: s.send(Packet({'Command': "msg_level " + args.msglevel, 'ID': packet_id + "_1"})()) if args.name is not None: s.send(Packet({'Command': "name " + args.name, 'ID': packet_id + "_2"})()) if args.contest is not None: s.send(Packet({'Command': "select_contest " + args.contest, 'ID': packet_id + "_3"})()) except IOError as e: console.write("\033[31;1mexception while reconnecting: " + str(e) + "\033[0m\n") queue.append((reconnect, ())) def handle_socket( handle, events ): if events & select.EPOLLIN: events &= ~select.EPOLLIN try: parser.add(s.recv(4096, socket.MSG_DONTWAIT)) queue.append((handle_parser, ())) except IOError as e: console.write("\033[31;1mlost connection to testsys: recv: " + str(e) + "\033[0m\n", force=True) queue.append((reconnect, ())) if events & select.EPOLLERR: events &= ~select.EPOLLERR console.write("\033[31;1mlost connection to testsys: err\033[0m\n", force=True) queue.append((reconnect, ())) if events & select.EPOLLHUP: events &= ~select.EPOLLHUP console.write("\033[31;1mlost connection to testsys: hup\033[0m\n", force=True) queue.append((reconnect, ())) if events != 0: console.write("ERROR: cannot handle event %d (h=%d)\n" % (events, handle), force=True) count = 200 def handle_stdin( handle, events ): if events & select.EPOLLIN: events &= ~select.EPOLLIN console.add(sys.stdin.read(4096)) if events != 0: console.write("ERROR: cannot handle event %d (h=%d)\n" % (events, handle), force=True) count = 4 def handle_parser(): global is_reconnect for packet in parser(): # console.write("[debug] work with packet %s\n" % str(packet)) if 'Log' in packet: console.write(packet['Log'] + '\n') if 'Message' in packet: if packet['Message'] is not None: message = packet['Message'] if message[-1] != '\n': message += '\n' console.write(message) else: console.write('\033[31;1mERROR: “Message” field exists in packet but is None.\033[0m\n') if 'Chat' in packet: console.write('\033[1m' + packet['Chat'] + '\033[0m\n') if 'ID' in packet: if console.state is Console.LOCK and console.value == packet['ID']: if is_reconnect: console.unlock ('', wait=False) is_reconnect = False else: console.unlock('... press any key ...') # TODO: check for ignored keys is_reconnect = False action = {s.fileno() : handle_socket, sys.stdin.fileno(): handle_stdin} parser = PacketParser() count = 0 while True: queue = [] # console.write("[debug] ready to poll\n", force=True) try: for handle, events in poll.poll(): # console.write("[debug] poll: handle=%d, event_mask=%d (IN=%d,OUT=%d,ERR=%d)\n" % (handle, events, select.EPOLLIN, select.EPOLLOUT, select.EPOLLERR), force=True) if handle in action: queue.append((action[handle], (handle, events))) else: console.write("ERROR: cannot handle %d\n" % handle, force=True) except IOError as e: console.write("\033[31;1mERROR: " + str(e) + "\033[0m\n") queue.append((reconnect, ())) except KeyboardInterrupt: console.lock("terminated by KeyboardInterrupt", "never") print(""); break except Exception: console.write("\033[31;1mERROR: " + str(e) + "\033[0m\n") break for f, p in queue: # console.write("[debug] next action\n") if isinstance(p, tuple): f(*p) else: f(p) # console.write("[debug] out of actions\n") if count > 10: break termios.tcsetattr(fd, termios.TCSADRAIN, tty_attr_old) #estSysConsoleTerminal::TestSysConsoleTerminal() : in(), console(), debug(), socket(), mutex(0), lock(), waitID(""), history(), counter(0) #oid TestSysConsoleTerminal::exec( const String &name, String const &msglevel, const String &server, const String &keyFileName ) # # File debugFile = File::openWrite("console_term.log"); # TerminalDevice debugConsole = TerminalDevice(console); # debug.addOutput(debugFile, DEBUG_FULLLOGGING); # debug.addOutput(debugConsole, DEBUG_EXTRADEBUG); # # bool exitFlag = false; # # tamias::MethodThread<TestSysConsoleTerminal> thread = tamias::MethodThread<TestSysConsoleTerminal>(this, &TestSysConsoleTerminal::secondLoop); # while (true) { # try { # thread.create(); # dts::PacketParser parser; # while (true) { # while (parser.packetReady()) # { # Packet packet = parser.nextPacket(); # for (sizetype i = 0; i < packet.values().size(); i++) # { # String packetName = escapeString(String::fromUtf8(Packet::ibm8662utf8(packet.values()[i].first))); # String packetValue = escapeString(String::fromUtf8(Packet::ibm8662utf8(packet.values()[i].second))); # debug.output(DEBUG_FULLLOGGING, " “%s” = “%s”") << packetName << packetValue; # } # for (sizetype i = 0; i < packet.values().size(); i++) # { # String packetName = String::fromUtf8(Packet::ibm8662utf8(packet.values()[i].first)); # String packetValue = String::fromUtf8(Packet::ibm8662utf8(packet.values()[i].second)); # else if (packetName != "") # { # // lock.lock(); # debug.output(DEBUG_DIAGNOSTIC, "unknow field in packet: “%s” -> “%s”") << packetName << packetValue; // TODO: semaphore!!! # // lock.unlock(); # } # } # } # } # // lock.lock(); # thread.cancel(); # // lock.unlock(); # console.setInput(" *** disconnected ***", 0); # socket.disconnect(); # if (exitFlag) # break; # tamias::Thread::sleep(1); # } catch (tamias::Exception &e ) { # debug.output(DEBUG_INFORMATION, "exception!"); # socket.disconnect(); # } # } # # #tring TestSysConsoleTerminal::readCommand() # # Vector <Pair <String, sizetype> > currentHistory; # for (sizetype i = 0; i < history.size(); i++) # currentHistory.pushBack(makePair(history[i], history[i].length())); # sizetype historyIndex = currentHistory.size(); # currentHistory.pushBack(makePair(String(""), 0)); #/ String command = ""; #/ sizetype cursor = 0; # while (true) # { # String &command = currentHistory[historyIndex].first; # sizetype &cursor = currentHistory[historyIndex].second; # // lock.lock(); # console.setInput("> " + command, cursor + 2); # // lock.unlock(); # int key = in.nextKey(); # switch (key) # { # case TerminalReader::KEY_UP: # if (historyIndex > 0) # historyIndex--; # break; # case TerminalReader::KEY_DOWN: # if (historyIndex < currentHistory.size() - 1) # historyIndex++; # break; # case TerminalReader::KEY_LEFT: # case TerminalReader::KEY_RIGHT: # case TerminalReader::KEY_HOME: # case TerminalReader::KEY_END: # case TerminalReader::KEY_BACKSPACE: # case TerminalReader::KEY_BACKSPACE2: # case TerminalReader::KEY_TAB: # case TerminalReader::KEY_DELETE: # case TerminalReader::KEY_ENTER: # // lock.lock(); # console.setInput("", 0); # console.output("> " + command + "\n"); # // lock.unlock(); # return command; # default: # } # } # # #oid TestSysConsoleTerminal::secondLoop() # # while (true) # { # String command = readCommand(); #/ bool was = false; #/ for (sizetype i = 0; i < history.size(); i++) #/ if (history[i] == command) #/ was = true, i = history.size(); #/ TODO: better ignore dups # if (history.size() == 0 || history[history.size() - 1] != command) # history.pushBack(command); #/ TODO: while (history.size() >= 100) #/ { #/ for (sizetype i = 1; i < history.size(); i++) #/ history[i - 1] = history[i]; #/ history.popBack(); #/ } # // lock.lock(); # console.setInput(" --- waiting for testsys outcome ---", 0); # // lock.unlock(); # // TODO: exit by Ctrl+D # String id = tamias::Format::intToString(counter++); # while (id.length() < 8) id = '0' + id; id = "id_" + id; # Packet packet; # packet.addValue("Command", Packet::utf82ibm866(command.toUtf8())); # packet.addValue("ID", id.toUtf8()); # // lock.lock(); # waitID = id; # mutex.set(1); # socket.write(packet.result()); # // lock.unlock(); # mutex.wait(1); # // lock.lock(); # console.setInput(" --- press any key ---", 0); # // lock.unlock(); # in.nextKey(); # mutex.set(0, true); # } # # # #namespace dts { # class TestSysConsoleTerminal { # public: # TestSysConsoleTerminal(); # ~TestSysConsoleTerminal(); # void exec( const tamias::String &name, tamias::String const &msglevel, const tamias::String &server, const tamias::String &keyFileName ); # # private: # enum DebugLevel { # DEBUG_NOLOGGING = 0, # DEBUG_FATAL = 1, # DEBUG_ERROR = 2, # DEBUG_WARNING = 3, # DEBUG_INFORMATION = 4, # DEBUG_DIAGNOSTIC = 5, # DEBUG_DEBUG = 6, # DEBUG_EXTRADEBUG = 7, # DEBUG_FULLLOGGING = 8 # }; # # burunduk3::TerminalReader in; # burunduk3::TerminalConsole console; # tamias::wtf::DebugOutput debug; # tamias::TcpClientSocket socket; # tamias::Mutex mutex; # tamias::Semaphore lock; # tamias::String waitID; # tamias::Vector <tamias::String> history; # tamias::Vector <tamias::String> testsysCommands; # int counter; # # tamias::String readCommand(); # tamias::String tabCompletion( const tamias::String &prefix ); # void secondLoop(); # # TestSysConsoleTerminal( const TestSysConsoleTerminal &termninal ); # TestSysConsoleTerminal& operator = ( const TestSysConsoleTerminal &terminal ); # }; #}
gpl-3.0
7,367,696,288,431,108,000
36.91619
174
0.550035
false
deculler/DataScienceTableDemos
CalGrads/timetable.py
1
7684
from datascience import Table import numpy as np from scipy.interpolate import UnivariateSpline class TimeTable(Table): """Table with a designated column as a sequence of times in the first column.""" def __init__(self, *args, time_column = 'Year'): Table.__init__(self, *args) self.time_column = time_column def clone_bare(self): return TimeTable(time_column = self.time_column) def clone_time(self): return self.clone_bare().with_column(self.time_column, self[self.time_column]) @classmethod def from_table(cls, tbl, time_col): ttbl = cls(time_column = time_col) for label in tbl.labels: ttbl[label] = tbl[label] return ttbl def __get_attr__(self, name): def wrapper(*args, **kwargs): # Wrap superclass method to coerce result back to TimeTable tbl = self.name(*args, **kwargs) if isinstance(tbl, Table) and self.time_column in tbl.labels: return TimeTable.from_table(tbl, self.time_column) else: return tbl print("Get TimeTable Attr", name) if hasattr(Table, name): return wrapper else: raise AttributeError @classmethod def by_time(cls, tbl, time_col, category_col, collect_col, collect=sum): """Construct a time table by aggregating rows of each category by year.""" tbl_by_year = tbl.select([category_col, time_col, collect_col]).pivot(category_col, time_col, collect_col, collect=collect) return cls(tbl_by_year.labels, time_column=time_col).append(tbl_by_year) @property def categories(self): return [label for label in self.labels if label != self.time_column] # TimeTable methods utilizing time_column def order_cols(self): """Create a TimeTable with categories ordered by the values in last row.""" def col_key(label): return self.row(self.num_rows-1)[self.labels.index(label)] order = sorted(self.categories, key=col_key, reverse=True) tbl = self.copy() for label in order: tbl.move_to_end(label) return tbl def oplot(self, **kwargs): return self.order_cols().plot(self.time_column, **kwargs) def top(self, n): """Create a new TimeTable containing the n largest columns.""" ttbl = self.order_cols() return ttbl.select(ttbl.labels[0:n+1]) def after(self, timeval): return self.where(self[self.time_column] >= timeval) def sum_rows(self): """Sum the rows in a TimeTable besides the time column.""" tbl = self.drop(self.time_column) return [sum(row) for row in tbl.rows] def apply_cols(self, fun): """Apply a function to the non-time columns of TimeTable.""" return Table().with_columns([(lbl, fun(self[lbl])) for lbl in self.categories]) def apply_all(self, fun): ttbl = TimeTable(time_column = self.time_column) for lbl in self.labels: if lbl == self.time_column: ttbl[lbl] = self[self.time_column] else: ttbl[lbl] = self.apply(fun, lbl) return ttbl def ratio(self, tbl_denom): """Create ratio of a TimeTable to a matching one.""" rtbl = TimeTable(time_column = self.time_column).with_column(self.time_column, self[self.time_column]) for label in self.categories: rtbl[label] = self[label] / tbl_denom[label] return rtbl def normalize(self, col_label): """Normalize each column of a timetable by a particular one""" rtbl = TimeTable(time_column = self.time_column).with_column(self.time_column, self[self.time_column]) for label in self.categories: rtbl[label] = self[label] / self[col_label] return rtbl def delta(self): """Construct a TimeTableable of successive differences down each non-time column.""" delta_tbl = self.clone_bare() delta_tbl[self.time_column] = self[self.time_column][1:] for col in self.categories: delta_tbl[col] = self[col][1:] - self[col][:-1] return delta_tbl def fill(self, interval=1): times = [t for t in np.arange(self[self.time_column][0], self[self.time_column][-1] + interval, interval)] ftbl = TimeTable(time_column = self.time_column).with_column(self.time_column, times) for col in self.categories: spl = UnivariateSpline(self[self.time_column], self[col]) ftbl[col] = spl(times) return ftbl def interp(self, interval=1): times = [t for t in np.arange(self[self.time_column][0], self[self.time_column][-1] + interval, interval)] ftbl = TimeTable(time_column = self.time_column).with_column(self.time_column, times) for col in self.categories: ftbl[col] = np.interp(times, self[self.time_column], self[col]) return ftbl def rel_delta(self): """Construct a TimeTableable of successive differences down each non-time column.""" delta_tbl = self.clone_bare() delta_tbl[self.time_column] = self[self.time_column][1:] time_delta = self[self.time_column][1:] - self[self.time_column][:-1] for col in self.categories: delta_tbl[col] = (1+(self[col][1:] - self[col][:-1])/self[col][:-1])/time_delta return delta_tbl def norm_by_row(self, base_row=0): """Normalize columns of a TimeTable by a row""" normed_tbl = self.clone_time() for label in self.categories: normed_tbl[label] = self[label]/self[label][base_row] return normed_tbl def norm_by_time(self, time): return self.norm_by_row(np.where(self[self.time_column] == time)[0][0]) def sum_cols(self): """Sum the columns of TimeTable.""" csum = 0 for c in self.categories: csum += self[c] return csum def fraction_cols(self): """Convert each column to a fraction by row.""" total = self.sum_cols() ftbl = self.clone_time() for lbl in self.categories: ftbl[lbl] = self[lbl]/total return ftbl def forecast_table(self, past, ahead, inc=1): """Project a TimeTable forward. inc must match the interval""" last_time = self[self.time_column][-1] past_times = self[self.time_column][-past-1:-1] fore_time = np.arange(last_time + inc, last_time + inc + ahead, inc) def project(lbl): m, b = np.polyfit(past_times, self[lbl][-past-1:-1], 1) return [m*time + b for time in fore_time] xtbl = Table().with_columns([(self.time_column, fore_time)] + [(label, project(label)) for label in self.categories]) return self.copy().append(xtbl) def extend_table(self, ahead, inc=1): """Project a TimeTable forward from last interval. inc must match the interval""" last_time = self[self.time_column][-1] fore_time = np.arange(last_time + inc, last_time + inc + ahead, inc) def project(lbl): b = self[lbl][-1] m = self[lbl][-1] - self[lbl][-2] return [m*(time+1)*inc + b for time in range(ahead)] xtbl = Table().with_columns([(self.time_column, fore_time)] + [(label, project(label)) for label in self.categories]) return self.copy().append(xtbl)
bsd-2-clause
1,163,413,192,845,707,000
40.989071
125
0.587585
false
elliotthill/django-oscar
oscar/apps/order/migrations/0019_auto__chg_field_order_billing_address__chg_field_order_user__chg_field.py
1
37418
# -*- coding: utf-8 -*- from south.utils import datetime_utils as datetime from south.db import db from south.v2 import SchemaMigration from django.db import models from oscar.core.compat import AUTH_USER_MODEL, AUTH_USER_MODEL_NAME class Migration(SchemaMigration): def forwards(self, orm): # Changing field 'Order.billing_address' db.alter_column(u'order_order', 'billing_address_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['order.BillingAddress'], null=True, on_delete=models.SET_NULL)) # Changing field 'Order.user' db.alter_column(u'order_order', 'user_id', self.gf('django.db.models.fields.related.ForeignKey')(null=True, on_delete=models.SET_NULL, to=orm[AUTH_USER_MODEL])) # Changing field 'Order.shipping_address' db.alter_column(u'order_order', 'shipping_address_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['order.ShippingAddress'], null=True, on_delete=models.SET_NULL)) def backwards(self, orm): # Changing field 'Order.billing_address' db.alter_column(u'order_order', 'billing_address_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['order.BillingAddress'], null=True)) # Changing field 'Order.user' db.alter_column(u'order_order', 'user_id', self.gf('django.db.models.fields.related.ForeignKey')(null=True, to=orm[AUTH_USER_MODEL])) # Changing field 'Order.shipping_address' db.alter_column(u'order_order', 'shipping_address_id', self.gf('django.db.models.fields.related.ForeignKey')(to=orm['order.ShippingAddress'], null=True)) models = { u'address.country': { 'Meta': {'ordering': "('-display_order', 'name')", 'object_name': 'Country'}, 'display_order': ('django.db.models.fields.PositiveSmallIntegerField', [], {'default': '0', 'db_index': 'True'}), 'is_shipping_country': ('django.db.models.fields.BooleanField', [], {'default': 'False', 'db_index': 'True'}), 'iso_3166_1_a2': ('django.db.models.fields.CharField', [], {'max_length': '2', 'primary_key': 'True'}), 'iso_3166_1_a3': ('django.db.models.fields.CharField', [], {'max_length': '3', 'null': 'True', 'db_index': 'True'}), 'iso_3166_1_numeric': ('django.db.models.fields.PositiveSmallIntegerField', [], {'null': 'True', 'db_index': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'printable_name': ('django.db.models.fields.CharField', [], {'max_length': '128'}) }, u'auth.group': { 'Meta': {'object_name': 'Group'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '80'}), 'permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}) }, u'auth.permission': { 'Meta': {'ordering': "(u'content_type__app_label', u'content_type__model', u'codename')", 'unique_together': "((u'content_type', u'codename'),)", 'object_name': 'Permission'}, 'codename': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'content_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['contenttypes.ContentType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}) }, AUTH_USER_MODEL: { 'Meta': {'object_name': AUTH_USER_MODEL_NAME}, 'date_joined': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'email': ('django.db.models.fields.EmailField', [], {'max_length': '75', 'blank': 'True'}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'groups': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'related_name': "u'user_set'", 'blank': 'True', 'to': u"orm['auth.Group']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'is_staff': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_superuser': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'last_login': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'password': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'user_permissions': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'related_name': "u'user_set'", 'blank': 'True', 'to': u"orm['auth.Permission']"}), 'username': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '30'}) }, u'catalogue.attributeentity': { 'Meta': {'object_name': 'AttributeEntity'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'slug': ('django.db.models.fields.SlugField', [], {'max_length': '255', 'blank': 'True'}), 'type': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'entities'", 'to': u"orm['catalogue.AttributeEntityType']"}) }, u'catalogue.attributeentitytype': { 'Meta': {'object_name': 'AttributeEntityType'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'slug': ('django.db.models.fields.SlugField', [], {'max_length': '255', 'blank': 'True'}) }, u'catalogue.attributeoption': { 'Meta': {'object_name': 'AttributeOption'}, 'group': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'options'", 'to': u"orm['catalogue.AttributeOptionGroup']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'option': ('django.db.models.fields.CharField', [], {'max_length': '255'}) }, u'catalogue.attributeoptiongroup': { 'Meta': {'object_name': 'AttributeOptionGroup'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128'}) }, u'catalogue.category': { 'Meta': {'ordering': "['full_name']", 'object_name': 'Category'}, 'depth': ('django.db.models.fields.PositiveIntegerField', [], {}), 'description': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'full_name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'db_index': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'image': ('django.db.models.fields.files.ImageField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'db_index': 'True'}), 'numchild': ('django.db.models.fields.PositiveIntegerField', [], {'default': '0'}), 'path': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '255'}), 'slug': ('django.db.models.fields.SlugField', [], {'max_length': '255'}) }, u'catalogue.option': { 'Meta': {'object_name': 'Option'}, 'code': ('django.db.models.fields.SlugField', [], {'unique': 'True', 'max_length': '128'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'type': ('django.db.models.fields.CharField', [], {'default': "'Required'", 'max_length': '128'}) }, u'catalogue.product': { 'Meta': {'ordering': "['-date_created']", 'object_name': 'Product'}, 'attributes': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['catalogue.ProductAttribute']", 'through': u"orm['catalogue.ProductAttributeValue']", 'symmetrical': 'False'}), 'categories': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['catalogue.Category']", 'through': u"orm['catalogue.ProductCategory']", 'symmetrical': 'False'}), 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'date_updated': ('django.db.models.fields.DateTimeField', [], {'auto_now': 'True', 'db_index': 'True', 'blank': 'True'}), 'description': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_discountable': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'parent': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'variants'", 'null': 'True', 'to': u"orm['catalogue.Product']"}), 'product_class': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'products'", 'null': 'True', 'to': u"orm['catalogue.ProductClass']"}), 'product_options': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['catalogue.Option']", 'symmetrical': 'False', 'blank': 'True'}), 'rating': ('django.db.models.fields.FloatField', [], {'null': 'True'}), 'recommended_products': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['catalogue.Product']", 'symmetrical': 'False', 'through': u"orm['catalogue.ProductRecommendation']", 'blank': 'True'}), 'related_products': ('django.db.models.fields.related.ManyToManyField', [], {'symmetrical': 'False', 'related_name': "'relations'", 'blank': 'True', 'to': u"orm['catalogue.Product']"}), 'score': ('django.db.models.fields.FloatField', [], {'default': '0.0', 'db_index': 'True'}), 'slug': ('django.db.models.fields.SlugField', [], {'max_length': '255'}), 'status': ('django.db.models.fields.CharField', [], {'db_index': 'True', 'max_length': '128', 'null': 'True', 'blank': 'True'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'upc': ('django.db.models.fields.CharField', [], {'max_length': '64', 'unique': 'True', 'null': 'True', 'blank': 'True'}) }, u'catalogue.productattribute': { 'Meta': {'ordering': "['code']", 'object_name': 'ProductAttribute'}, 'code': ('django.db.models.fields.SlugField', [], {'max_length': '128'}), 'entity_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.AttributeEntityType']", 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'option_group': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.AttributeOptionGroup']", 'null': 'True', 'blank': 'True'}), 'product_class': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'attributes'", 'null': 'True', 'to': u"orm['catalogue.ProductClass']"}), 'required': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'type': ('django.db.models.fields.CharField', [], {'default': "'text'", 'max_length': '20'}) }, u'catalogue.productattributevalue': { 'Meta': {'object_name': 'ProductAttributeValue'}, 'attribute': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.ProductAttribute']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'product': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'attribute_values'", 'to': u"orm['catalogue.Product']"}), 'value_boolean': ('django.db.models.fields.NullBooleanField', [], {'null': 'True', 'blank': 'True'}), 'value_date': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}), 'value_entity': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.AttributeEntity']", 'null': 'True', 'blank': 'True'}), 'value_file': ('django.db.models.fields.files.FileField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'value_float': ('django.db.models.fields.FloatField', [], {'null': 'True', 'blank': 'True'}), 'value_image': ('django.db.models.fields.files.ImageField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'value_integer': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}), 'value_option': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.AttributeOption']", 'null': 'True', 'blank': 'True'}), 'value_richtext': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'value_text': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}) }, u'catalogue.productcategory': { 'Meta': {'ordering': "['-is_canonical']", 'object_name': 'ProductCategory'}, 'category': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.Category']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_canonical': ('django.db.models.fields.BooleanField', [], {'default': 'False', 'db_index': 'True'}), 'product': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.Product']"}) }, u'catalogue.productclass': { 'Meta': {'ordering': "['name']", 'object_name': 'ProductClass'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'options': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['catalogue.Option']", 'symmetrical': 'False', 'blank': 'True'}), 'requires_shipping': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'slug': ('django.db.models.fields.SlugField', [], {'unique': 'True', 'max_length': '128'}), 'track_stock': ('django.db.models.fields.BooleanField', [], {'default': 'True'}) }, u'catalogue.productrecommendation': { 'Meta': {'object_name': 'ProductRecommendation'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'primary': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'primary_recommendations'", 'to': u"orm['catalogue.Product']"}), 'ranking': ('django.db.models.fields.PositiveSmallIntegerField', [], {'default': '0'}), 'recommendation': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.Product']"}) }, u'contenttypes.contenttype': { 'Meta': {'ordering': "('name',)", 'unique_together': "(('app_label', 'model'),)", 'object_name': 'ContentType', 'db_table': "'django_content_type'"}, 'app_label': ('django.db.models.fields.CharField', [], {'max_length': '100'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'model': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '100'}) }, u'customer.communicationeventtype': { 'Meta': {'object_name': 'CommunicationEventType'}, 'category': ('django.db.models.fields.CharField', [], {'default': "u'Order related'", 'max_length': '255'}), 'code': ('django.db.models.fields.SlugField', [], {'max_length': '128'}), 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'date_updated': ('django.db.models.fields.DateTimeField', [], {'auto_now': 'True', 'blank': 'True'}), 'email_body_html_template': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'email_body_template': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'email_subject_template': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'sms_template': ('django.db.models.fields.CharField', [], {'max_length': '170', 'null': 'True', 'blank': 'True'}) }, u'order.billingaddress': { 'Meta': {'object_name': 'BillingAddress'}, 'country': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['address.Country']"}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'blank': 'True'}), 'line1': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'line2': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'line3': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'line4': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'postcode': ('oscar.models.fields.UppercaseCharField', [], {'max_length': '64', 'null': 'True', 'blank': 'True'}), 'search_text': ('django.db.models.fields.CharField', [], {'max_length': '1000'}), 'state': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '64', 'null': 'True', 'blank': 'True'}) }, u'order.communicationevent': { 'Meta': {'ordering': "['-date_created']", 'object_name': 'CommunicationEvent'}, 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'event_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['customer.CommunicationEventType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'communication_events'", 'to': u"orm['order.Order']"}) }, u'order.line': { 'Meta': {'object_name': 'Line'}, 'est_dispatch_date': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'line_price_before_discounts_excl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'line_price_before_discounts_incl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'line_price_excl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'line_price_incl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'lines'", 'to': u"orm['order.Order']"}), 'partner': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'order_lines'", 'null': 'True', 'on_delete': 'models.SET_NULL', 'to': u"orm['partner.Partner']"}), 'partner_line_notes': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'partner_line_reference': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'blank': 'True'}), 'partner_name': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'partner_sku': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'product': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['catalogue.Product']", 'null': 'True', 'on_delete': 'models.SET_NULL', 'blank': 'True'}), 'quantity': ('django.db.models.fields.PositiveIntegerField', [], {'default': '1'}), 'status': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'stockrecord': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['partner.StockRecord']", 'null': 'True', 'on_delete': 'models.SET_NULL', 'blank': 'True'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'unit_cost_price': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'unit_price_excl_tax': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'unit_price_incl_tax': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'unit_retail_price': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'upc': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'blank': 'True'}) }, u'order.lineattribute': { 'Meta': {'object_name': 'LineAttribute'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'line': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'attributes'", 'to': u"orm['order.Line']"}), 'option': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'line_attributes'", 'null': 'True', 'on_delete': 'models.SET_NULL', 'to': u"orm['catalogue.Option']"}), 'type': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'value': ('django.db.models.fields.CharField', [], {'max_length': '255'}) }, u'order.lineprice': { 'Meta': {'ordering': "('id',)", 'object_name': 'LinePrice'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'line': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'prices'", 'to': u"orm['order.Line']"}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'line_prices'", 'to': u"orm['order.Order']"}), 'price_excl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'price_incl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'quantity': ('django.db.models.fields.PositiveIntegerField', [], {'default': '1'}), 'shipping_excl_tax': ('django.db.models.fields.DecimalField', [], {'default': '0', 'max_digits': '12', 'decimal_places': '2'}), 'shipping_incl_tax': ('django.db.models.fields.DecimalField', [], {'default': '0', 'max_digits': '12', 'decimal_places': '2'}) }, u'order.order': { 'Meta': {'ordering': "['-date_placed']", 'object_name': 'Order'}, 'basket_id': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True', 'blank': 'True'}), 'billing_address': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['order.BillingAddress']", 'null': 'True', 'on_delete': 'models.SET_NULL', 'blank': 'True'}), 'currency': ('django.db.models.fields.CharField', [], {'default': "'GBP'", 'max_length': '12'}), 'date_placed': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'db_index': 'True', 'blank': 'True'}), 'guest_email': ('django.db.models.fields.EmailField', [], {'max_length': '75', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'number': ('django.db.models.fields.CharField', [], {'max_length': '128', 'db_index': 'True'}), 'shipping_address': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['order.ShippingAddress']", 'null': 'True', 'on_delete': 'models.SET_NULL', 'blank': 'True'}), 'shipping_code': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '128', 'blank': 'True'}), 'shipping_excl_tax': ('django.db.models.fields.DecimalField', [], {'default': '0', 'max_digits': '12', 'decimal_places': '2'}), 'shipping_incl_tax': ('django.db.models.fields.DecimalField', [], {'default': '0', 'max_digits': '12', 'decimal_places': '2'}), 'shipping_method': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'blank': 'True'}), 'site': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['sites.Site']"}), 'status': ('django.db.models.fields.CharField', [], {'max_length': '100', 'null': 'True', 'blank': 'True'}), 'total_excl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'total_incl_tax': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'user': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'orders'", 'null': 'True', 'on_delete': 'models.SET_NULL', 'to': u"orm['{}']".format(AUTH_USER_MODEL)}) }, u'order.orderdiscount': { 'Meta': {'object_name': 'OrderDiscount'}, 'amount': ('django.db.models.fields.DecimalField', [], {'default': '0', 'max_digits': '12', 'decimal_places': '2'}), 'category': ('django.db.models.fields.CharField', [], {'default': "'Basket'", 'max_length': '64'}), 'frequency': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'message': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'offer_id': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True', 'blank': 'True'}), 'offer_name': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'db_index': 'True'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'discounts'", 'to': u"orm['order.Order']"}), 'voucher_code': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'db_index': 'True'}), 'voucher_id': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True', 'blank': 'True'}) }, u'order.ordernote': { 'Meta': {'object_name': 'OrderNote'}, 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'date_updated': ('django.db.models.fields.DateTimeField', [], {'auto_now': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'message': ('django.db.models.fields.TextField', [], {}), 'note_type': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'notes'", 'to': u"orm['order.Order']"}), 'user': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['{}']".format(AUTH_USER_MODEL), 'null': 'True'}) }, u'order.paymentevent': { 'Meta': {'ordering': "['-date_created']", 'object_name': 'PaymentEvent'}, 'amount': ('django.db.models.fields.DecimalField', [], {'max_digits': '12', 'decimal_places': '2'}), 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'event_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['order.PaymentEventType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'lines': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['order.Line']", 'through': u"orm['order.PaymentEventQuantity']", 'symmetrical': 'False'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'payment_events'", 'to': u"orm['order.Order']"}), 'reference': ('django.db.models.fields.CharField', [], {'max_length': '128', 'blank': 'True'}), 'shipping_event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'payment_events'", 'null': 'True', 'to': u"orm['order.ShippingEvent']"}) }, u'order.paymenteventquantity': { 'Meta': {'object_name': 'PaymentEventQuantity'}, 'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'line_quantities'", 'to': u"orm['order.PaymentEvent']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'line': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'payment_event_quantities'", 'to': u"orm['order.Line']"}), 'quantity': ('django.db.models.fields.PositiveIntegerField', [], {}) }, u'order.paymenteventtype': { 'Meta': {'ordering': "('name',)", 'object_name': 'PaymentEventType'}, 'code': ('django.db.models.fields.SlugField', [], {'unique': 'True', 'max_length': '128'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '128'}) }, u'order.shippingaddress': { 'Meta': {'object_name': 'ShippingAddress'}, 'country': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['address.Country']"}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '255', 'blank': 'True'}), 'line1': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'line2': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'line3': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'line4': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'notes': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'phone_number': ('oscar.models.fields.PhoneNumberField', [], {'max_length': '128', 'blank': 'True'}), 'postcode': ('oscar.models.fields.UppercaseCharField', [], {'max_length': '64', 'null': 'True', 'blank': 'True'}), 'search_text': ('django.db.models.fields.CharField', [], {'max_length': '1000'}), 'state': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '64', 'null': 'True', 'blank': 'True'}) }, u'order.shippingevent': { 'Meta': {'ordering': "['-date_created']", 'object_name': 'ShippingEvent'}, 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'event_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['order.ShippingEventType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'lines': ('django.db.models.fields.related.ManyToManyField', [], {'related_name': "'shipping_events'", 'symmetrical': 'False', 'through': u"orm['order.ShippingEventQuantity']", 'to': u"orm['order.Line']"}), 'notes': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'order': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'shipping_events'", 'to': u"orm['order.Order']"}) }, u'order.shippingeventquantity': { 'Meta': {'object_name': 'ShippingEventQuantity'}, 'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'line_quantities'", 'to': u"orm['order.ShippingEvent']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'line': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'shipping_event_quantities'", 'to': u"orm['order.Line']"}), 'quantity': ('django.db.models.fields.PositiveIntegerField', [], {}) }, u'order.shippingeventtype': { 'Meta': {'ordering': "('name',)", 'object_name': 'ShippingEventType'}, 'code': ('django.db.models.fields.SlugField', [], {'unique': 'True', 'max_length': '128'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '255'}) }, u'partner.partner': { 'Meta': {'object_name': 'Partner'}, 'code': ('django.db.models.fields.SlugField', [], {'unique': 'True', 'max_length': '128'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '128', 'null': 'True', 'blank': 'True'}), 'users': ('django.db.models.fields.related.ManyToManyField', [], {'blank': 'True', 'related_name': "'partners'", 'null': 'True', 'symmetrical': 'False', 'to': u"orm['{}']".format(AUTH_USER_MODEL)}) }, u'partner.stockrecord': { 'Meta': {'unique_together': "(('partner', 'partner_sku'),)", 'object_name': 'StockRecord'}, 'cost_price': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'date_created': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'date_updated': ('django.db.models.fields.DateTimeField', [], {'auto_now': 'True', 'db_index': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'low_stock_threshold': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True', 'blank': 'True'}), 'num_allocated': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}), 'num_in_stock': ('django.db.models.fields.PositiveIntegerField', [], {'null': 'True', 'blank': 'True'}), 'partner': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'stockrecords'", 'to': u"orm['partner.Partner']"}), 'partner_sku': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'price_currency': ('django.db.models.fields.CharField', [], {'default': "'GBP'", 'max_length': '12'}), 'price_excl_tax': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'price_retail': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '12', 'decimal_places': '2', 'blank': 'True'}), 'product': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'stockrecords'", 'to': u"orm['catalogue.Product']"}) }, u'sites.site': { 'Meta': {'ordering': "(u'domain',)", 'object_name': 'Site', 'db_table': "u'django_site'"}, 'domain': ('django.db.models.fields.CharField', [], {'max_length': '100'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}) } } complete_apps = ['order']
bsd-3-clause
-2,684,955,528,462,074,000
90.043796
224
0.562163
false
danger89/stanford-parser-clientserver
stanford_server.py
1
3076
#!/usr/bin/env jython # -*- coding: utf-8 -*- # Copyright 2014 by Melroy van den Berg """ Stanford Parser Server running on localhost Requirements ------------ - Jython >= 2.7 (http://www.jython.org/downloads.html) - Pyro4 (https://github.com/irmen/Pyro4) """ __author__ = "Melroy van den Berg <[email protected]>" __version__ = "0.1" import socket from select import cpython_compatible_select as select import sys import Pyro4.core import Pyro4.naming from stanford_interface import StanfordParser PYRO_NAME = 'stanford.server' Pyro4.config.SERVERTYPE="thread" # Thread pool based #Pyro4.config.SERVERTYPE="multiplex" # Select/poll based hostname="localhost" #socket.gethostname() class StanfordHelpParser(object): """ Helper class around the StanfordParser class """ def __init__(self): """ Setup the Stanford Parser """ # Jar file should be set inside the stanford_lib.py or add it manually to the class path self.parser = StanfordParser(parser_file='./englishPCFG.ser.gz') def parse(self, wordList): """ Parse the word list """ sentenceObject = self.parser.parse_wordlist(wordList) return str(sentenceObject.get_parse()) print("initializing services... servertype=%s" % Pyro4.config.SERVERTYPE) # start a name server (only with a broadcast server when NOT running on localhost) nameserverUri, nameserverDaemon, broadcastServer = Pyro4.naming.startNS(host=hostname) print("got a Nameserver, uri=%s" % nameserverUri) print("ns daemon location string=%s" % nameserverDaemon.locationStr) print("ns daemon sockets=%s" % nameserverDaemon.sockets) if broadcastServer: print("bc server socket=%s (fileno %d)" % (broadcastServer.sock, broadcastServer.fileno())) # create a Pyro daemon pyrodaemon=Pyro4.core.Daemon(host=hostname) print("daemon location string=%s" % pyrodaemon.locationStr) print("daemon sockets=%s" % pyrodaemon.sockets) # register a server object with the daemon serveruri=pyrodaemon.register(StanfordHelpParser()) print("server uri=%s" % serveruri) # register it with the embedded nameserver directly nameserverDaemon.nameserver.register(PYRO_NAME,serveruri) print("Stanford Server is running...") # below is our custom event loop. while True: # create sets of the socket objects we will be waiting on # (a set provides fast lookup compared to a list) nameserverSockets = set(nameserverDaemon.sockets) pyroSockets = set(pyrodaemon.sockets) rs=[] if broadcastServer: rs=[broadcastServer] # only the broadcast server is directly usable as a select() object rs.extend(nameserverSockets) rs.extend(pyroSockets) rs,_,_ = select(rs,[],[],2) eventsForNameserver=[] eventsForDaemon=[] for s in rs: if s is broadcastServer: broadcastServer.processRequest() elif s in nameserverSockets: eventsForNameserver.append(s) elif s in pyroSockets: eventsForDaemon.append(s) if eventsForNameserver: nameserverDaemon.events(eventsForNameserver) if eventsForDaemon: pyrodaemon.events(eventsForDaemon) nameserverDaemon.close() if broadcastServer: broadcastServer.close() pyrodaemon.close()
apache-2.0
5,812,613,648,402,874,000
29.156863
92
0.75065
false