content
stringlengths 7
928k
| avg_line_length
float64 3.5
33.8k
| max_line_length
int64 6
139k
| alphanum_fraction
float64 0.08
0.96
| licenses
sequence | repository_name
stringlengths 7
104
| path
stringlengths 4
230
| size
int64 7
928k
| lang
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|
# --------------------------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# --------------------------------------------------------------------------------------------
import io
from queue import Queue
import time
import unittest
import unittest.mock as mock
from ossdbtoolsservice.hosting.json_rpc_server import (
JSONRPCServer,
IncomingMessageConfiguration,
NotificationContext, RequestContext
)
from ossdbtoolsservice.hosting.json_message import JSONRPCMessage, JSONRPCMessageType
from ossdbtoolsservice.hosting.json_reader import JSONRPCReader
from ossdbtoolsservice.hosting.json_writer import JSONRPCWriter
import tests.utils as utils
class JSONRPCServerTests(unittest.TestCase):
def test_handler_init(self):
# If: I create a Handler class
handler = JSONRPCServer.Handler('class', 'handler')
# Then: The values should be available
self.assertEqual(handler.class_, 'class')
self.assertEqual(handler.handler, 'handler')
def test_server_init(self):
# Setup: Create objects to init the server with
input_stream = io.BytesIO()
output_stream = io.BytesIO()
logger = utils.get_mock_logger()
# If: I create a server
server = JSONRPCServer(input_stream, output_stream, logger=logger)
# Then: The state should be initialized as defined
self.assertIsInstance(server.writer, JSONRPCWriter)
self.assertIsInstance(server.reader, JSONRPCReader)
self.assertIs(server._logger, logger)
self.assertEqual(server._version, '0')
self.assertFalse(server._stop_requested)
# ... The output queue should be empty
self.assertIsInstance(server._output_queue, Queue)
self.assertTrue(server._output_queue.all_tasks_done)
self.assertDictEqual(server._notification_handlers, {})
self.assertListEqual(server._shutdown_handlers, [])
# ... The threads shouldn't be assigned yet
self.assertIsNone(server._output_consumer)
self.assertIsNone(server._input_consumer)
# ... The built-in handlers should be assigned
self.assertTrue('echo' in server._request_handlers)
self.assertIsNotNone(server._request_handlers['echo'])
self.assertTrue('version' in server._request_handlers)
self.assertIsNotNone(server._request_handlers['version'].handler)
self.assertTrue('shutdown' in server._request_handlers)
self.assertIsNotNone(server._request_handlers['shutdown'].handler)
self.assertTrue('exit' in server._request_handlers)
self.assertIsNotNone(server._request_handlers['exit'].handler)
def test_add_shutdown_handler(self):
# If: I add a shutdown handler
handler = mock.MagicMock()
server = JSONRPCServer(None, None)
server.add_shutdown_handler(handler)
# Then: The shutdown handlers should contain the handler
self.assertTrue(handler in server._shutdown_handlers)
def test_set_request_handler(self):
# If: I add a request handler
params = IncomingMessageConfiguration('test/test', int)
handler = mock.MagicMock()
server = JSONRPCServer(None, None)
server.set_request_handler(params, handler)
# Then: The request handler should contain the handler
self.assertTrue(params.method in server._request_handlers)
self.assertIsNotNone(server._request_handlers[params.method])
self.assertIs(server._request_handlers[params.method].class_, int)
self.assertIs(server._request_handlers[params.method].handler, handler)
def test_set_notification_handler(self):
# If: I add a notification handler
params = IncomingMessageConfiguration('test/test', int)
handler = mock.MagicMock()
server = JSONRPCServer(None, None)
server.set_notification_handler(params, handler)
# Then: The request handler should contain the handler
self.assertTrue(params.method in server._notification_handlers)
self.assertIsNotNone(server._notification_handlers[params.method])
self.assertIs(server._notification_handlers[params.method].class_, int)
self.assertIs(server._notification_handlers[params.method].handler, handler)
# BUILT-IN HANDLER TESTS ###############################################
@staticmethod
def test_echo_request():
# If: I send a request for an echo
rc = utils.MockRequestContext()
params = {}
JSONRPCServer._handle_echo_request(rc, params)
# Then: The params should have been echoed back
rc.send_response.assert_called_once_with(params)
rc.send_notification.assert_not_called()
rc.send_error.assert_not_called()
@staticmethod
def test_version_request():
# If: I send a request for the version
rc = utils.MockRequestContext()
server = JSONRPCServer(None, None)
server._handle_version_request(rc, None)
# Then: I should get a response
rc.send_response.assert_called_once_with(server._version)
rc.send_error.assert_not_called()
rc.send_notification.assert_not_called()
def test_shutdown_request(self):
# If: I send a request for the service to shutdown
rc = utils.MockRequestContext()
handler = mock.MagicMock()
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server.add_shutdown_handler(handler)
server._handle_shutdown_request(rc, None)
# Then:
# ... The server should be shutting down
self.assertTrue(server._stop_requested)
# ... The shutdown handler should be called
handler.assert_called_once()
# RequestContext TESTS #################################################
def test_request_context_init_test(self):
# If: I create a request context
queue = Queue()
message = JSONRPCMessage.from_dictionary({'id': '123', 'method': 'test/text/', 'params': {}})
rc = RequestContext(message, queue)
# Then: The internal state should be set up correctly
self.assertIs(rc._message, message)
self.assertIs(rc._queue, queue)
def test_request_context_send_response(self):
# Setup: Create a request context
queue = Queue()
in_message = JSONRPCMessage.from_dictionary({'id': '123', 'method': 'test/text/', 'params': {}})
rc = RequestContext(in_message, queue)
# If: I send a response via the response handler
params = {}
rc.send_response(params)
# Then:
# ... There should be a message in the outbound queue
self.assertTrue(queue.not_empty)
out_message = queue.get_nowait()
self.assertIsInstance(out_message, JSONRPCMessage)
# .. The message must be a response with the proper id
self.assertEqual(out_message.message_type, JSONRPCMessageType.ResponseSuccess)
self.assertEqual(out_message.message_id, '123')
self.assertEqual(out_message.message_result, params)
def test_request_context_send_notification(self):
# Setup: Create a request context
queue = Queue()
in_message = JSONRPCMessage.from_dictionary({'id': '123', 'method': 'test/text/', 'params': {}})
rc = RequestContext(in_message, queue)
# If: I send a notification
params = {}
method = 'test/test'
rc.send_notification(method, params)
# Then:
# ... There should be a message in the outbound queue
self.assertTrue(queue.not_empty)
out_message = queue.get_nowait()
self.assertIsInstance(out_message, JSONRPCMessage)
# .. The message must be a response with the proper id
self.assertEqual(out_message.message_type, JSONRPCMessageType.Notification)
self.assertIsNone(out_message.message_id)
self.assertEqual(out_message.message_params, params)
def test_request_context_send_error(self):
# Setup: Create a request context
queue = Queue()
in_message = JSONRPCMessage.from_dictionary({'id': '123', 'method': 'test/text/', 'params': {}})
rc = RequestContext(in_message, queue)
# If: I send an error
params = {}
rc.send_error(params)
# Then:
# ... There should be a message in the outbound queue
self.assertTrue(queue.not_empty)
out_message = queue.get_nowait()
self.assertIsInstance(out_message, JSONRPCMessage)
# .. The message must be a response with the proper id
self.assertEqual(out_message.message_type, JSONRPCMessageType.ResponseError)
self.assertEqual(out_message.message_id, '123')
self.assertIsInstance(out_message.message_error, dict)
self.assertIs(out_message.message_error['message'], params)
# DISPATCHER TESTS #####################################################
@staticmethod
def test_dispatch_response_success():
# TODO: Replace with robust logic once response routing is implemented
# If: I dispatch a response message
message = JSONRPCMessage.create_response('123', {})
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server._dispatch_message(message)
# Then: Nothing should have happened
@staticmethod
def test_dispatch_response_error():
# TODO: Replace with robust logic once error routing is implemented
# If: I dispatch an error message
message = JSONRPCMessage.create_error('123', 0, message='', data={})
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server._dispatch_message(message)
# Then: Nothing should have happened
@staticmethod
def test_dispatch_invalid():
# If: I dispatch an invalid message
message = JSONRPCMessage('invalidType')
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server._dispatch_message(message)
# Then: Nothing should have happened
@staticmethod
def test_dispatch_request_no_handler():
# If: I dispatch a message that has no handler
logger = utils.get_mock_logger()
message = JSONRPCMessage.create_request('123', 'non_existent', {})
server = JSONRPCServer(None, None, logger=logger)
server._dispatch_message(message)
# Then:
# ... Nothing should have happened
# TODO: Capture that an error was sent
# ... A warning should have been logged
logger.warn.assert_called_once()
def test_dispatch_request_none_class(self):
# Setup: Create a server with a single handler that has none for the deserialization class
config = IncomingMessageConfiguration('test/test', None)
handler = mock.MagicMock()
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server.set_request_handler(config, handler)
# If: I dispatch a message that has none set for the deserialization class
params = {}
message = JSONRPCMessage.create_request('123', 'test/test', params)
server._dispatch_message(message)
# Then:
# ... The handler should have been called
handler.assert_called_once()
# ... The parameters to the handler should have been a request context and params
self.assertIsInstance(handler.mock_calls[0][1][0], RequestContext)
self.assertIs(handler.mock_calls[0][1][0]._queue, server._output_queue)
self.assertIs(handler.mock_calls[0][1][0]._message, message)
self.assertIs(handler.mock_calls[0][1][1], params)
def test_dispatch_request_normal(self):
# Setup: Create a server with a single handler that has none for the deserialization class
config = IncomingMessageConfiguration('test/test', _TestParams)
handler = mock.MagicMock()
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server.set_request_handler(config, handler)
# If: I dispatch a message that has none set for the deserialization class
params = {}
message = JSONRPCMessage.create_request('123', 'test/test', params)
server._dispatch_message(message)
# Then:
# ... The handler should have been called
handler.assert_called_once()
# ... The parameters to the handler should have been a request context and params
self.assertIsInstance(handler.mock_calls[0][1][0], RequestContext)
self.assertIs(handler.mock_calls[0][1][0]._queue, server._output_queue)
self.assertIs(handler.mock_calls[0][1][0]._message, message)
self.assertIsInstance(handler.mock_calls[0][1][1], _TestParams)
@staticmethod
def test_dispatch_notification_no_handler():
# If: I dispatch a message that has no handler
logger = utils.get_mock_logger()
message = JSONRPCMessage.create_notification('non_existent', {})
server = JSONRPCServer(None, None, logger=logger)
server._dispatch_message(message)
# Then:
# ... Nothing should have happened
# TODO: Capture that an error was sent
# ... A warning should have been logged
logger.warn.assert_called_once()
def test_dispatch_notification_none_class(self):
# Setup: Create a server with a single handler that has none for the deserialization class
config = IncomingMessageConfiguration('test/test', None)
handler = mock.MagicMock()
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server.set_notification_handler(config, handler)
# If: I dispatch a message that has none set for the deserialization class
params = {}
message = JSONRPCMessage.create_notification('test/test', params)
server._dispatch_message(message)
# Then:
# ... The handler should have been called
handler.assert_called_once()
# ... The parameters to the handler should have been a request context and params
self.assertIsInstance(handler.mock_calls[0][1][0], NotificationContext)
self.assertIs(handler.mock_calls[0][1][0]._queue, server._output_queue)
self.assertIs(handler.mock_calls[0][1][1], params)
def test_dispatch_notification_normal(self):
# Setup: Create a server with a single handler that has none for the deserialization class
config = IncomingMessageConfiguration('test/test', _TestParams)
handler = mock.MagicMock()
server = JSONRPCServer(None, None, logger=utils.get_mock_logger())
server.set_notification_handler(config, handler)
# If: I dispatch a message that has none set for the deserialization class
params = {}
message = JSONRPCMessage.create_notification('test/test', params)
server._dispatch_message(message)
# Then:
# ... The handler should have been called
handler.assert_called_once()
# ... The parameters to the handler should have been a request context and params
self.assertIsInstance(handler.mock_calls[0][1][0], NotificationContext)
self.assertIs(handler.mock_calls[0][1][0]._queue, server._output_queue)
self.assertIsInstance(handler.mock_calls[0][1][1], _TestParams)
# RequestContext TESTS #################################################
def test_notification_context_init_test(self):
# If: I create a notification context
queue = Queue()
nc = NotificationContext(queue)
# Then: The internal state should be set up correctly
self.assertIs(nc._queue, queue)
def test_notification_context_send(self):
# Setup: Create a request context
queue = Queue()
nc = NotificationContext(queue)
# If: I send a response via the response handler
method = 'test/test'
params = {}
nc.send_notification(method, params)
# Then:
# ... There should be a message in the outbound queue
self.assertTrue(queue.not_empty)
out_message = queue.get_nowait()
self.assertIsInstance(out_message, JSONRPCMessage)
# .. The message must be a response with the proper id
self.assertEqual(out_message.message_type, JSONRPCMessageType.Notification)
self.assertIsNone(out_message.message_id)
self.assertEqual(out_message.message_params, params)
self.assertEqual(out_message.message_method, method)
# END-TO-END TESTS #####################################################
def test_request_enqueued(self):
# Setup: Create empty io streams
input_stream = io.BytesIO()
output_stream = io.BytesIO()
# If: I submit an outbound request
test_client = JSONRPCServer(input_stream, output_stream)
test_client.send_request('test/test', {'test': 'test'})
# Then:
# ... There should be one request in the outbound queue
request = test_client._output_queue.get()
# ... The queued message should match the request we sent
self.assertEqual(request.message_method, 'test/test')
self.assertDictEqual(request.message_params, {'test': 'test'})
def test_notification_enqueued(self):
# Setup: Create empty io streams
input_stream = io.BytesIO()
output_stream = io.BytesIO()
# If: I submit an outbound request
test_client = JSONRPCServer(input_stream, output_stream)
test_client.send_notification('test/test', {'test': 'test'})
# Then:
# ... There should be one request in the outbound queue
request = test_client._output_queue.get()
# ... The queued message should match the request we sent
self.assertEqual(request.message_method, 'test/test')
self.assertDictEqual(request.message_params, {'test': 'test'})
def test_reads_message(self):
# Setup:
# ... Create an input stream with a single message
input_stream = io.BytesIO(b'Content-Length: 30\r\n\r\n{"method":"test", "params":{}}')
output_stream = io.BytesIO()
# ... Create a server that uses the input and output streams
server = JSONRPCServer(input_stream, output_stream, logger=utils.get_mock_logger())
# ... Patch the server to not dispatch a message
dispatch_mock = mock.MagicMock()
server._dispatch_message = dispatch_mock
# If: I start the server, run it for a bit, and stop it
# TODO: Remove explicit sleep and add spin-locks
server.start()
time.sleep(1)
server.stop()
server.wait_for_exit()
# Then: The dispatch method should have been called
expected_output = JSONRPCMessage.from_dictionary({"method": "test", "params": {}})
dispatch_mock.assert_called_once()
self.assertDictEqual(dispatch_mock.mock_calls[0][1][0].dictionary, expected_output.dictionary)
# Teardown: All background threads should be shut down.
self.assertFalse(server._input_consumer.isAlive())
self.assertFalse(server._output_consumer.isAlive())
def test_read_multiple_messages(self):
# Setup:
# ... Create an input stream with two messages
test_bytes = b'Content-Length: 30\r\n\r\n{"method":"test", "params":{}}'
input_stream = io.BytesIO(test_bytes + test_bytes)
output_stream = io.BytesIO()
# ... Create a server that uses the input and output streams
server = JSONRPCServer(input_stream, output_stream, logger=utils.get_mock_logger())
# ... Patch the server to not dispatch a message
dispatch_mock = mock.MagicMock()
server._dispatch_message = dispatch_mock
# If: I start the server, run it for a bit, and stop it
server.start()
time.sleep(1)
server.stop()
server.wait_for_exit()
# Then: The dispatch method should have been called twice
expected_output = JSONRPCMessage.from_dictionary({"method": "test", "params": {}})
self.assertEqual(len(dispatch_mock.mock_calls), 2)
self.assertDictEqual(dispatch_mock.mock_calls[0][1][0].dictionary, expected_output.dictionary)
self.assertDictEqual(dispatch_mock.mock_calls[1][1][0].dictionary, expected_output.dictionary)
# Teardown: All background threads should be shut down.
self.assertFalse(server._input_consumer.isAlive())
self.assertFalse(server._output_consumer.isAlive())
class _TestParams:
@classmethod
def from_dict(cls, dictionary):
return _TestParams()
def __init__(self):
pass
if __name__ == '__main__':
unittest.main()
| 41.403162 | 104 | 0.664916 | [
"MIT"
] | DaeunYim/pgtoolsservice | tests/hosting/test_server.py | 20,950 | Python |
from setuptools import setup, find_packages
VERSION = '0.0.1'
DESCRIPTION = 'edu-lib'
LONG_DESCRIPTION = 'Libary zum erlernen der Grundstruktur.'
setup(
name="mylibrary",
version=VERSION,
author="Stephan Bökelmann",
author_email="[email protected]",
scripts=[],
description=DESCRIPTION,
long_description=LONG_DESCRIPTION,
packages=find_packages(),
install_requires=[],
url="",
keywords=['python', 'debugging'],
classifiers= [
"Intended Audience :: Education",
"Programming Language :: Python :: 3",
"Operating System :: POSIX",
]
)
| 26.192308 | 59 | 0.581498 | [
"MIT"
] | bjoekeldude/edu_python_mini_lib | setup.py | 682 | Python |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi SDK Generator. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union
from ... import _utilities, _tables
from . import outputs
from ._inputs import *
__all__ = ['Account']
class Account(pulumi.CustomResource):
def __init__(__self__,
resource_name: str,
opts: Optional[pulumi.ResourceOptions] = None,
account_name: Optional[pulumi.Input[str]] = None,
active_directories: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ActiveDirectoryArgs']]]]] = None,
location: Optional[pulumi.Input[str]] = None,
resource_group_name: Optional[pulumi.Input[str]] = None,
tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
__props__=None,
__name__=None,
__opts__=None):
"""
NetApp account resource
:param str resource_name: The name of the resource.
:param pulumi.ResourceOptions opts: Options for the resource.
:param pulumi.Input[str] account_name: The name of the NetApp account
:param pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['ActiveDirectoryArgs']]]] active_directories: Active Directories
:param pulumi.Input[str] location: Resource location
:param pulumi.Input[str] resource_group_name: The name of the resource group.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] tags: Resource tags
"""
if __name__ is not None:
warnings.warn("explicit use of __name__ is deprecated", DeprecationWarning)
resource_name = __name__
if __opts__ is not None:
warnings.warn("explicit use of __opts__ is deprecated, use 'opts' instead", DeprecationWarning)
opts = __opts__
if opts is None:
opts = pulumi.ResourceOptions()
if not isinstance(opts, pulumi.ResourceOptions):
raise TypeError('Expected resource options to be a ResourceOptions instance')
if opts.version is None:
opts.version = _utilities.get_version()
if opts.id is None:
if __props__ is not None:
raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource')
__props__ = dict()
__props__['account_name'] = account_name
__props__['active_directories'] = active_directories
__props__['location'] = location
if resource_group_name is None and not opts.urn:
raise TypeError("Missing required property 'resource_group_name'")
__props__['resource_group_name'] = resource_group_name
__props__['tags'] = tags
__props__['name'] = None
__props__['provisioning_state'] = None
__props__['type'] = None
alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:netapp/v20200901:Account"), pulumi.Alias(type_="azure-native:netapp:Account"), pulumi.Alias(type_="azure-nextgen:netapp:Account"), pulumi.Alias(type_="azure-native:netapp/latest:Account"), pulumi.Alias(type_="azure-nextgen:netapp/latest:Account"), pulumi.Alias(type_="azure-native:netapp/v20170815:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20170815:Account"), pulumi.Alias(type_="azure-native:netapp/v20190501:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20190501:Account"), pulumi.Alias(type_="azure-native:netapp/v20190601:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20190601:Account"), pulumi.Alias(type_="azure-native:netapp/v20190701:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20190701:Account"), pulumi.Alias(type_="azure-native:netapp/v20190801:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20190801:Account"), pulumi.Alias(type_="azure-native:netapp/v20191001:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20191001:Account"), pulumi.Alias(type_="azure-native:netapp/v20191101:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20191101:Account"), pulumi.Alias(type_="azure-native:netapp/v20200201:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200201:Account"), pulumi.Alias(type_="azure-native:netapp/v20200301:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200301:Account"), pulumi.Alias(type_="azure-native:netapp/v20200501:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200501:Account"), pulumi.Alias(type_="azure-native:netapp/v20200601:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200601:Account"), pulumi.Alias(type_="azure-native:netapp/v20200701:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200701:Account"), pulumi.Alias(type_="azure-native:netapp/v20200801:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20200801:Account"), pulumi.Alias(type_="azure-native:netapp/v20201101:Account"), pulumi.Alias(type_="azure-nextgen:netapp/v20201101:Account")])
opts = pulumi.ResourceOptions.merge(opts, alias_opts)
super(Account, __self__).__init__(
'azure-native:netapp/v20200901:Account',
resource_name,
__props__,
opts)
@staticmethod
def get(resource_name: str,
id: pulumi.Input[str],
opts: Optional[pulumi.ResourceOptions] = None) -> 'Account':
"""
Get an existing Account resource's state with the given name, id, and optional extra
properties used to qualify the lookup.
:param str resource_name: The unique name of the resulting resource.
:param pulumi.Input[str] id: The unique provider ID of the resource to lookup.
:param pulumi.ResourceOptions opts: Options for the resource.
"""
opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id))
__props__ = dict()
__props__["active_directories"] = None
__props__["location"] = None
__props__["name"] = None
__props__["provisioning_state"] = None
__props__["tags"] = None
__props__["type"] = None
return Account(resource_name, opts=opts, __props__=__props__)
@property
@pulumi.getter(name="activeDirectories")
def active_directories(self) -> pulumi.Output[Optional[Sequence['outputs.ActiveDirectoryResponse']]]:
"""
Active Directories
"""
return pulumi.get(self, "active_directories")
@property
@pulumi.getter
def location(self) -> pulumi.Output[str]:
"""
Resource location
"""
return pulumi.get(self, "location")
@property
@pulumi.getter
def name(self) -> pulumi.Output[str]:
"""
Resource name
"""
return pulumi.get(self, "name")
@property
@pulumi.getter(name="provisioningState")
def provisioning_state(self) -> pulumi.Output[str]:
"""
Azure lifecycle management
"""
return pulumi.get(self, "provisioning_state")
@property
@pulumi.getter
def tags(self) -> pulumi.Output[Optional[Mapping[str, str]]]:
"""
Resource tags
"""
return pulumi.get(self, "tags")
@property
@pulumi.getter
def type(self) -> pulumi.Output[str]:
"""
Resource type
"""
return pulumi.get(self, "type")
def translate_output_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
def translate_input_property(self, prop):
return _tables.SNAKE_TO_CAMEL_CASE_TABLE.get(prop) or prop
| 51.065789 | 2,057 | 0.675213 | [
"Apache-2.0"
] | pulumi-bot/pulumi-azure-native | sdk/python/pulumi_azure_native/netapp/v20200901/account.py | 7,762 | Python |
from django.test import TestCase
from blog.models import Entry
class EntryModelTest(TestCase):
def test_string_representation(self):
entry = Entry(title="My entry title")
self.assertEqual(str(entry), entry.title)
def test_verbose_name_plural(self):
self.assertEqual(str(Entry._meta.verbose_name_plural), "entries")
| 27 | 73 | 0.735043 | [
"BSD-2-Clause"
] | elinguiuriel/djangoTDD | code/blog/tests/test_models.py | 351 | Python |
import os
import logging
from django.conf import settings
from django.utils import translation
from django.utils.translation import gettext_lazy as _
from django.db import transaction
from django.core.files.base import ContentFile
from celery.exceptions import SoftTimeLimitExceeded
from froide.celery import app as celery_app
from froide.publicbody.models import PublicBody
from froide.upload.models import Upload
from .models import FoiRequest, FoiMessage, FoiAttachment, FoiProject
from .foi_mail import _process_mail, _fetch_mail
from .notifications import send_classification_reminder
logger = logging.getLogger(__name__)
@celery_app.task(
name="froide.foirequest.tasks.process_mail", acks_late=True, time_limit=60
)
def process_mail(*args, **kwargs):
translation.activate(settings.LANGUAGE_CODE)
with transaction.atomic():
_process_mail(*args, **kwargs)
@celery_app.task(name="froide.foirequest.tasks.fetch_mail", expires=60)
def fetch_mail():
for mail_uid, rfc_data in _fetch_mail():
process_mail.delay(rfc_data, mail_uid=mail_uid)
@celery_app.task
def detect_overdue():
translation.activate(settings.LANGUAGE_CODE)
for foirequest in FoiRequest.objects.get_to_be_overdue():
foirequest.set_overdue()
@celery_app.task
def detect_asleep():
translation.activate(settings.LANGUAGE_CODE)
for foirequest in FoiRequest.objects.get_to_be_asleep():
foirequest.set_asleep()
@celery_app.task
def classification_reminder():
translation.activate(settings.LANGUAGE_CODE)
for foirequest in FoiRequest.objects.get_unclassified():
send_classification_reminder(foirequest)
@celery_app.task
def check_delivery_status(message_id, count=None, extended=False):
try:
message = FoiMessage.objects.get(id=message_id)
except FoiMessage.DoesNotExist:
return
message.check_delivery_status(count=count, extended=extended)
@celery_app.task
def create_project_requests(project_id, publicbody_ids, **kwargs):
for seq, pb_id in enumerate(publicbody_ids):
create_project_request.delay(project_id, pb_id, sequence=seq, **kwargs)
@celery_app.task
def create_project_request(project_id, publicbody_id, sequence=0, **kwargs):
from .services import CreateRequestFromProjectService
try:
project = FoiProject.objects.get(id=project_id)
except FoiProject.DoesNotExist:
# project does not exist anymore?
return
try:
pb = PublicBody.objects.get(id=publicbody_id)
except PublicBody.DoesNotExist:
# pb was deleted?
return
kwargs.update(
{
"project": project,
"publicbody": pb,
"subject": project.title,
"user": project.user,
"body": project.description,
"public": project.public,
"reference": project.reference,
"tags": [t.name for t in project.tags.all()],
"project_order": sequence,
}
)
service = CreateRequestFromProjectService(kwargs)
foirequest = service.execute()
if project.request_count == project.foirequest_set.all().count():
project.status = FoiProject.STATUS_READY
project.save()
return foirequest.pk
@celery_app.task(name="froide.foirequest.tasks.convert_attachment_task", time_limit=60)
def convert_attachment_task(instance_id):
try:
att = FoiAttachment.objects.get(pk=instance_id)
except FoiAttachment.DoesNotExist:
return
if att.can_convert_to_pdf():
return convert_attachment(att)
def ocr_pdf_attachment(att):
if att.converted:
ocred_att = att.converted
else:
name, ext = os.path.splitext(att.name)
name = _("{name}_ocr{ext}").format(name=name, ext=".pdf")
ocred_att = FoiAttachment.objects.create(
name=name,
belongs_to=att.belongs_to,
approved=False,
filetype="application/pdf",
is_converted=True,
can_approve=att.can_approve,
)
att.converted = ocred_att
att.can_approve = False
att.approved = False
att.save()
ocr_pdf_task.delay(
att.pk,
ocred_att.pk,
)
def convert_attachment(att):
from filingcabinet.pdf_utils import convert_to_pdf
output_bytes = convert_to_pdf(
att.file.path,
binary_name=settings.FROIDE_CONFIG.get("doc_conversion_binary"),
construct_call=settings.FROIDE_CONFIG.get("doc_conversion_call_func"),
)
if output_bytes is None:
return
if att.converted:
new_att = att.converted
else:
name, ext = os.path.splitext(att.name)
name = _("{name}_converted{ext}").format(name=name, ext=".pdf")
new_att = FoiAttachment(
name=name,
belongs_to=att.belongs_to,
approved=False,
filetype="application/pdf",
is_converted=True,
can_approve=att.can_approve,
)
new_file = ContentFile(output_bytes)
new_att.size = new_file.size
new_att.file.save(new_att.name, new_file)
new_att.save()
att.converted = new_att
att.can_approve = False
att.approved = False
att.save()
@celery_app.task(
name="froide.foirequest.tasks.convert_images_to_pdf_task",
time_limit=60 * 5,
soft_time_limit=60 * 4,
)
def convert_images_to_pdf_task(att_ids, target_id, instructions, can_approve=True):
from filingcabinet.pdf_utils import convert_images_to_ocred_pdf
att_qs = FoiAttachment.objects.filter(id__in=att_ids)
att_map = {a.id: a for a in att_qs}
atts = [att_map[a_id] for a_id in att_ids]
try:
target = FoiAttachment.objects.get(id=target_id)
except FoiAttachment.DoesNotExist:
return
paths = [a.file.path for a in atts]
try:
pdf_bytes = convert_images_to_ocred_pdf(paths, instructions=instructions)
except SoftTimeLimitExceeded:
pdf_bytes = None
if pdf_bytes is None:
att_qs.update(can_approve=can_approve)
target.delete()
return
new_file = ContentFile(pdf_bytes)
target.size = new_file.size
target.file.save(target.name, new_file)
target.save()
@celery_app.task(
name="froide.foirequest.tasks.ocr_pdf_task",
time_limit=60 * 5,
soft_time_limit=60 * 4,
)
def ocr_pdf_task(att_id, target_id, can_approve=True):
from filingcabinet.pdf_utils import run_ocr
try:
attachment = FoiAttachment.objects.get(pk=att_id)
except FoiAttachment.DoesNotExist:
return
try:
target = FoiAttachment.objects.get(pk=target_id)
except FoiAttachment.DoesNotExist:
return
try:
pdf_bytes = run_ocr(
attachment.file.path,
language=settings.TESSERACT_LANGUAGE
if settings.TESSERACT_LANGUAGE
else settings.LANGUAGE_CODE,
timeout=180,
)
except SoftTimeLimitExceeded:
pdf_bytes = None
if pdf_bytes is None:
attachment.can_approve = can_approve
attachment.save()
target.delete()
return
new_file = ContentFile(pdf_bytes)
target.size = new_file.size
target.file.save(target.name, new_file)
target.save()
@celery_app.task(
name="froide.foirequest.tasks.redact_attachment_task",
time_limit=60 * 6,
soft_time_limit=60 * 5,
)
def redact_attachment_task(att_id, target_id, instructions):
from filingcabinet.pdf_utils import run_ocr
from froide.helper.redaction import redact_file
try:
attachment = FoiAttachment.objects.get(pk=att_id)
except FoiAttachment.DoesNotExist:
return
if att_id != target_id:
try:
target = FoiAttachment.objects.get(pk=target_id)
except FoiAttachment.DoesNotExist:
return
else:
target = attachment
logger.info("Trying redaction of %s", attachment.id)
try:
pdf_bytes = redact_file(attachment.file, instructions)
except Exception:
logger.error("PDF redaction error", exc_info=True)
pdf_bytes = None
if pdf_bytes is None:
logger.info("Redaction failed %s", attachment.id)
# Redaction has failed, remove empty attachment
if attachment.redacted:
attachment.redacted = None
if attachment.is_redacted:
attachment.approved = True
attachment.can_approve = True
attachment.pending = False
attachment.save()
if not target.file:
target.delete()
return
logger.info("Redaction successful %s", attachment.id)
pdf_file = ContentFile(pdf_bytes)
target.size = pdf_file.size
target.file.save(target.name, pdf_file, save=False)
logger.info("Trying OCR %s", target.id)
try:
pdf_bytes = run_ocr(
target.file.path,
language=settings.TESSERACT_LANGUAGE
if settings.TESSERACT_LANGUAGE
else settings.LANGUAGE_CODE,
timeout=60 * 4,
)
except SoftTimeLimitExceeded:
pdf_bytes = None
if pdf_bytes is not None:
logger.info("OCR successful %s", target.id)
pdf_file = ContentFile(pdf_bytes)
target.size = pdf_file.size
target.file.save(target.name, pdf_file, save=False)
else:
logger.info("OCR failed %s", target.id)
target.can_approve = True
target.pending = False
target.approve_and_save()
FoiAttachment.attachment_published.send(sender=target, user=None)
@celery_app.task(name="froide.foirequest.tasks.move_upload_to_attachment")
def move_upload_to_attachment(att_id, upload_id):
try:
att = FoiAttachment.objects.get(pk=att_id)
except FoiAttachment.DoesNotExist:
return
try:
upload = Upload.objects.get(pk=upload_id)
except FoiAttachment.DoesNotExist:
return
file = upload.get_file()
if file:
att.pending = False
att.file.save(att.name, file, save=True)
upload.finish()
upload.delete()
if att.can_convert_to_pdf():
convert_attachment_task.delay(att.id)
| 28.307263 | 87 | 0.677719 | [
"MIT"
] | MrKrisKrisu/froide | froide/foirequest/tasks.py | 10,134 | Python |
from __future__ import absolute_import
import pytest
from changes.expanders.commands import CommandsExpander
from changes.testutils import TestCase
class CommandsExpanderTest(TestCase):
def setUp(self):
super(CommandsExpanderTest, self).setUp()
self.project = self.create_project()
def get_expander(self, data):
return CommandsExpander(self.project, data)
def test_validate(self):
with pytest.raises(AssertionError):
self.get_expander({}).validate()
self.get_expander({'commands': []}).validate()
def test_expand(self):
project = self.create_project()
build = self.create_build(project)
job = self.create_job(build)
results = list(self.get_expander({'commands': [
{'script': 'echo 1'},
{'script': 'echo 2', 'label': 'foo'}
]}).expand(job=job, max_executors=10))
assert len(results) == 2
assert results[0].label == 'echo 1'
assert len(results[0].commands) == 1
assert results[0].commands[0].label == 'echo 1'
assert results[0].commands[0].script == 'echo 1'
assert results[1].label == 'foo'
assert len(results[1].commands) == 1
assert results[1].commands[0].label == 'foo'
assert results[1].commands[0].script == 'echo 2'
| 31.785714 | 56 | 0.626966 | [
"Apache-2.0"
] | dropbox/changes | tests/changes/expanders/test_commands.py | 1,335 | Python |
"""Parse Warren2020 fluxes.
Fluxes from https://zenodo.org/record/3952926 (DOI:10.5281/zenodo.3952926)
See https://arxiv.org/abs/1902.01340 and https://arxiv.org/abs/1912.03328
for description of the models.
"""
import h5py
from sntools.formats import gamma, get_starttime, get_endtime
flux = {}
def parse_input(input, inflv, starttime, endtime):
"""Read simulations data from input file.
Arguments:
input -- prefix of file containing neutrino fluxes
inflv -- neutrino flavor to consider
starttime -- start time set by user via command line option (or None)
endtime -- end time set by user via command line option (or None)
"""
f = h5py.File(input, 'r')
for (t, r) in f['sim_data']['shock_radius']:
if r > 1:
tbounce = t * 1000 # convert to ms
break
starttime = get_starttime(starttime, 1000 * f['sim_data']['shock_radius'][0][0] - tbounce)
endtime = get_endtime(endtime, 1000 * f['sim_data']['shock_radius'][-1][0] - tbounce)
# Save flux data to dictionary to look up in nu_emission() below
global flux
flux = {}
path = {'e': 'nue_data', 'eb': 'nuae_data', 'x': 'nux_data', 'xb': 'nux_data'}[inflv]
for i, (t, lum) in enumerate(f[path]['lum']):
t = 1000 * t - tbounce # convert to time post-bounce in ms
if (t < starttime - 30) or (t > endtime + 30):
# Ignore data outside of the requested time span.
continue
lum *= 1e51 * 624.151 # convert from 10^51 erg/s to MeV/ms
mean_e = f[path]['avg_energy'][i][1]
mean_e_sq = f[path]['rms_energy'][i][1]**2
flux[t] = (mean_e, mean_e_sq, lum)
f.close()
return (starttime, endtime, sorted(flux.keys()))
def prepare_evt_gen(binned_t):
global flux
gamma.flux = flux
gamma.prepare_evt_gen(binned_t)
flux = gamma.flux
def nu_emission(eNu, time):
gamma.flux = flux
return gamma.nu_emission(eNu, time)
| 30.578125 | 94 | 0.633112 | [
"BSD-3-Clause"
] | arfon/sntools | sntools/formats/warren2020.py | 1,957 | Python |
# Copyright 2016 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""
Sample Google App Engine application that demonstrates using the Users API
For more information about App Engine, see README.md under /appengine.
"""
# [START all]
from google.appengine.api import users
import webapp2
class MainPage(webapp2.RequestHandler):
def get(self):
user = users.get_current_user()
if user:
nickname = user.nickname()
logout_url = users.create_logout_url('/')
greeting = 'Welcome, {}! (<a href="{}">sign out</a>)'.format(
nickname, logout_url)
else:
login_url = users.create_login_url('/')
greeting = '<a href="{}">Sign in</a>'.format(login_url)
self.response.write(
'<html><body>{}</body></html>'.format(greeting))
class AdminPage(webapp2.RequestHandler):
def get(self):
user = users.get_current_user()
if user:
if users.is_current_user_admin():
self.response.write('You are an administrator.')
else:
self.response.write('You are not an administrator.')
else:
self.response.write('You are not logged in.')
app = webapp2.WSGIApplication([
('/', MainPage),
('/admin', AdminPage)
], debug=True)
# [END all]
| 30.278689 | 74 | 0.646995 | [
"Apache-2.0"
] | Acidburn0zzz/python-docs-samples | appengine/standard/users/main.py | 1,847 | Python |
"""
ASGI config for avocadobites project.
It exposes the ASGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/3.1/howto/deployment/asgi/
"""
import os
from django.core.asgi import get_asgi_application
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'avocadobites.settings')
application = get_asgi_application()
| 23.588235 | 78 | 0.790524 | [
"MIT"
] | sanjuop/PatrioticPictures | avocadobites/avocadobites/asgi.py | 401 | Python |
"""
eZmax API Definition
This API expose all the functionnalities for the eZmax and eZsign applications. # noqa: E501
The version of the OpenAPI document: 1.1.3
Contact: [email protected]
Generated by: https://openapi-generator.tech
"""
import sys
import unittest
import eZmaxApi
from eZmaxApi.model.ezsignformfield_response import EzsignformfieldResponse
globals()['EzsignformfieldResponse'] = EzsignformfieldResponse
from eZmaxApi.model.ezsignformfield_response_compound import EzsignformfieldResponseCompound
class TestEzsignformfieldResponseCompound(unittest.TestCase):
"""EzsignformfieldResponseCompound unit test stubs"""
def setUp(self):
pass
def tearDown(self):
pass
def testEzsignformfieldResponseCompound(self):
"""Test EzsignformfieldResponseCompound"""
# FIXME: construct object with mandatory attributes with example values
# model = EzsignformfieldResponseCompound() # noqa: E501
pass
if __name__ == '__main__':
unittest.main()
| 26.846154 | 97 | 0.747851 | [
"MIT"
] | eZmaxinc/eZmax-SDK-python | test/test_ezsignformfield_response_compound.py | 1,047 | Python |
import os
import shutil
import tempfile
from unittest import TestCase
from mock import patch
from regulations.apps import RegulationsConfig
class RegulationsConfigTests(TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def tearDown(self):
shutil.rmtree(self.tmpdir)
@patch('regulations.apps.get_app_template_dirs')
def test_precompute_custom_templates(self, get_app_template_dirs):
"""Verify that custom templates are found"""
get_app_template_dirs.return_value = [self.tmpdir]
open(os.path.join(self.tmpdir, '123-45-a.html'), 'w').close()
open(os.path.join(self.tmpdir, 'other.html'), 'w').close()
RegulationsConfig.precompute_custom_templates()
self.assertEqual(RegulationsConfig.custom_tpls['123-45-a'],
'regulations/custom_nodes/123-45-a.html')
self.assertEqual(RegulationsConfig.custom_tpls['other'],
'regulations/custom_nodes/other.html')
self.assertFalse('another' in RegulationsConfig.custom_tpls)
| 34.516129 | 70 | 0.695327 | [
"CC0-1.0"
] | CMSgov/cmcs-eregulations | regulations/tests/apps_tests.py | 1,070 | Python |
import random
from player import Player
from hand import Hand
class CPU(Player):
def __init__(self, name: str):
super().__init__(name)
self.hand = Hand()
def discard(self):
if(self.hand == None or len(self.hand) <= 0):
raise RuntimeError('No cards to discard')
return self.hand.pop(random.randrange(len(self.hand)))
def play(self, currentPlayPointLimit):
print('{0}\'s Hand: {1}'.format(self.name, str(self.playHand)))
if(self.playHand == None or len(self.playHand) <= 0):
raise RuntimeError('No play hand was created or it is empty')
playableCardIndexes = []
for i, card in enumerate(self.playHand):
if(card.valuePoints <= currentPlayPointLimit):
playableCardIndexes.append(i)
cardToPlayIndex = playableCardIndexes[random.randrange(len(playableCardIndexes))]
return self.playHand.pop(cardToPlayIndex) | 39.5 | 89 | 0.646624 | [
"MIT"
] | SudoSpartanDan/CribbagePythonGame | cpu.py | 948 | Python |
"""@package vc_updated
Functions to implement the updated Voce-Chaboche material model and measure its error.
"""
import numpy as np
import pandas as pd
from numdifftools import nd_algopy as nda
def uvc_return_mapping(x_sol, data, tol=1.0e-8, maximum_iterations=1000):
""" Implements the time integration of the updated Voce-Chaboche material model.
:param np.array x_sol: Updated Voce-Chaboche model parameters.
:param pd.DataFrame data: stress-strain data.
:param float tol: Local Newton tolerance.
:param int maximum_iterations: maximum iterations in local Newton procedure, raises RuntimeError if exceeded.
:return dict: History of: stress ('stress'), strain ('strain'), the total error ('error') calculated by the
updated Voce-Chaboche model, number of iterations for convergence at each increment ('num_its').
"""
if len(x_sol) < 8:
raise RuntimeError("No backstresses or using original V-C params.")
n_param_per_back = 2
n_basic_param = 6
# Get material properties
E = x_sol[0] * 1.0
sy_0 = x_sol[1] * 1.0
Q = x_sol[2] * 1.0
b = x_sol[3] * 1.0
D = x_sol[4] * 1.0
a = x_sol[5] * 1.0
# Set up backstresses
n_backstresses = int((len(x_sol) - n_basic_param) / n_param_per_back)
c_k = []
gamma_k = []
for i in range(0, n_backstresses):
c_k.append(x_sol[n_basic_param + n_param_per_back * i])
gamma_k.append(x_sol[n_basic_param + 1 + n_param_per_back * i])
# Initialize parameters
alpha_components = np.zeros(n_backstresses, dtype=object) # backstress components
strain = 0.
stress = 0.
ep_eq = 0. # equivalent plastic strain
error = 0. # error measure
sum_abs_de = 0. # total strain
stress_sim = 0.0
stress_test = 0.0
area_test = 0.0
stress_track = []
strain_track = []
strain_inc_track = []
iteration_track = []
loading = np.diff(data['e_true'])
for increment_number, strain_inc in enumerate(loading):
strain += strain_inc
alpha = np.sum(alpha_components)
yield_stress = sy_0 + Q * (1. - np.exp(-b * ep_eq)) - D * (1. - np.exp(-a * ep_eq))
trial_stress = stress + E * strain_inc
relative_stress = trial_stress - alpha
flow_dir = np.sign(relative_stress)
yield_condition = np.abs(relative_stress) - yield_stress
if yield_condition > tol:
is_converged = False
else:
is_converged = True
# For error
stress_sim_1 = stress_sim * 1.0
stress_test_1 = stress_test * 1.0
# Return mapping if plastic loading
ep_eq_init = ep_eq
alpha_init = alpha
consist_param = 0.
number_of_iterations = 0
while is_converged is False and number_of_iterations < maximum_iterations:
number_of_iterations += 1
# Isotropic hardening and isotropic modulus
yield_stress = sy_0 + Q * (1. - np.exp(-b * ep_eq)) - D * (1. - np.exp(-a * ep_eq))
iso_modulus = Q * b * np.exp(-b * ep_eq) - D * a * np.exp(-a * ep_eq)
# Kinematic hardening and kinematic modulus
alpha = 0.
kin_modulus = 0.
for i in range(0, n_backstresses):
e_k = np.exp(-gamma_k[i] * (ep_eq - ep_eq_init))
alpha += flow_dir * c_k[i] / gamma_k[i] + (alpha_components[i] - flow_dir * c_k[i] / gamma_k[i]) * e_k
kin_modulus += c_k[i] * e_k - flow_dir * gamma_k[i] * e_k * alpha_components[i]
delta_alpha = alpha - alpha_init
# Local Newton step
numerator = np.abs(relative_stress) - (consist_param * E + yield_stress + flow_dir * delta_alpha)
denominator = -(E + iso_modulus + kin_modulus)
consist_param = consist_param - numerator / denominator
ep_eq = ep_eq_init + consist_param
if np.abs(numerator) < tol:
is_converged = True
# Update the variables
stress = trial_stress - E * flow_dir * consist_param
for i in range(0, n_backstresses):
e_k = np.exp(-gamma_k[i] * (ep_eq - ep_eq_init))
alpha_components[i] = flow_dir * c_k[i] / gamma_k[i] \
+ (alpha_components[i] - flow_dir * c_k[i] / gamma_k[i]) * e_k
stress_track.append(stress)
strain_track.append(strain)
strain_inc_track.append(strain_inc)
iteration_track.append(number_of_iterations)
# Calculate the error
stress_sim = stress * 1.0
stress_test = data['Sigma_true'].iloc[increment_number + 1]
sum_abs_de += np.abs(strain_inc)
area_test += np.abs(strain_inc) * ((stress_test) ** 2 + (stress_test_1) ** 2) / 2.
error += np.abs(strain_inc) * ((stress_sim - stress_test) ** 2 + (stress_sim_1 - stress_test_1) ** 2) / 2.
if number_of_iterations >= maximum_iterations:
print ("Increment number = ", increment_number)
print ("Parameters = ", x_sol)
print ("Numerator = ", numerator)
raise RuntimeError('Return mapping did not converge in ' + str(maximum_iterations) + ' iterations.')
area = area_test / sum_abs_de
error = error / sum_abs_de
return {'stress': stress_track, 'strain': strain_track, 'error': error, 'num_its': iteration_track,
'area': area}
def sim_curve_uvc(x_sol, test_clean):
""" Returns the stress-strain approximation of the updated Voce-Chaboche material model to a given strain input.
:param np.array x_sol: Voce-Chaboche model parameters
:param DataFrame test_clean: stress-strain data
:return DataFrame: Voce-Chaboche approximation
The strain column in the DataFrame is labeled "e_true" and the stress column is labeled "Sigma_true".
"""
model_output = uvc_return_mapping(x_sol, test_clean)
strain = np.append([0.], model_output['strain'])
stress = np.append([0.], model_output['stress'])
sim_curve = pd.DataFrame(np.array([strain, stress]).transpose(), columns=['e_true', 'Sigma_true'])
return sim_curve
def error_single_test_uvc(x_sol, test_clean):
""" Returns the relative error between a test and its approximation using the updated Voce-Chaboche material model.
:param np.array x_sol: Voce-Chaboche model parameters
:param DataFrame test_clean: stress-strain data
:return float: relative error
The strain column in the DataFrame is labeled "e_true" and the stress column is labeled "Sigma_true".
"""
model_output = uvc_return_mapping(x_sol, test_clean)
return model_output['error']
def normalized_error_single_test_uvc(x_sol, test_clean):
""" Returns the error and the total area of a test and its approximation using the updated Voce-Chaboche
material model.
:param np.array x_sol: Voce-Chaboche model parameters
:param DataFrame test_clean: stress-strain data
:return list: (float) total error, (float) total area
The strain column in the DataFrame is labeled "e_true" and the stress column is labeled "Sigma_true".
"""
model_output = uvc_return_mapping(x_sol, test_clean)
return [model_output['error'], model_output['area']]
def calc_phi_total(x, data):
""" Returns the sum of the normalized relative error of the updated Voce-Chaboche material model given x.
:param np.array x: Updated Voce-Chaboche material model parameters.
:param list data: (pd.DataFrame) Stress-strain history for each test considered.
:return float: Normalized error value expressed as a percent (raw value * 100).
The normalized error is defined in de Sousa and Lignos (2017).
"""
error_total = 0.
area_total = 0.
for d in data:
error, area = normalized_error_single_test_uvc(x, d)
error_total += error
area_total += area
return np.sqrt(error_total / area_total) * 100.
def test_total_area(x, data):
""" Returns the total squared area underneath all the tests.
:param np.array x: Updated Voce-Chaboche material model parameters.
:param list data: (pd.DataFrame) Stress-strain history for each test considered.
:return float: Total squared area.
"""
area_total = 0.
for d in data:
_, area = normalized_error_single_test_uvc(x, d)
area_total += area
return area_total
def uvc_get_hessian(x, data):
""" Returns the Hessian of the material model error function for a given set of test data evaluated at x.
:param np.array x: Updated Voce-Chaboche material model parameters.
:param list data: (pd.DataFrame) Stress-strain history for each test considered.
:return np.array: Hessian matrix of the error function.
"""
def f(xi):
val = 0.
for d in data:
val += error_single_test_uvc(xi, d)
return val
hess_fun = nda.Hessian(f)
return hess_fun(x)
def uvc_consistency_metric(x_base, x_sample, data):
""" Returns the xi_2 consistency metric from de Sousa and Lignos 2019 using the updated Voce-Chaboche model.
:param np.array x_base: Updated Voce-Chaboche material model parameters from the base case.
:param np.array x_sample: Updated Voce-Chaboche material model parameters from the sample case.
:param list data: (pd.DataFrame) Stress-strain history for each test considered.
:return float: Increase in quadratic approximation from the base to the sample case.
"""
x_diff = x_sample - x_base
hess_base = uvc_get_hessian(x_base, data)
numerator = np.dot(x_diff, hess_base.dot(x_diff))
denominator = test_total_area(x_base, data)
return np.sqrt(numerator / denominator)
def uvc_tangent_modulus(x_sol, data, tol=1.0e-8, maximum_iterations=1000):
""" Returns the tangent modulus at each strain step.
:param np.array x_sol: Updated Voce-Chaboche model parameters.
:param pd.DataFrame data: stress-strain data.
:param float tol: Local Newton tolerance.
:param int maximum_iterations: maximum iterations in local Newton procedure, raises RuntimeError if exceeded.
:return np.ndarray: Tangent modulus array.
"""
if len(x_sol) < 8:
raise RuntimeError("No backstresses or using original V-C params.")
n_param_per_back = 2
n_basic_param = 6
# Get material properties
E = x_sol[0] * 1.0
sy_0 = x_sol[1] * 1.0
Q = x_sol[2] * 1.0
b = x_sol[3] * 1.0
D = x_sol[4] * 1.0
a = x_sol[5] * 1.0
# Set up backstresses
n_backstresses = int((len(x_sol) - n_basic_param) / n_param_per_back)
c_k = []
gamma_k = []
for i in range(0, n_backstresses):
c_k.append(x_sol[n_basic_param + n_param_per_back * i])
gamma_k.append(x_sol[n_basic_param + 1 + n_param_per_back * i])
# Initialize parameters
alpha_components = np.zeros(n_backstresses, dtype=object) # backstress components
strain = 0.
stress = 0.
ep_eq = 0. # equivalent plastic strain
stress_track = []
strain_track = []
strain_inc_track = []
iteration_track = []
tangent_track = []
loading = np.diff(data['e_true'])
for increment_number, strain_inc in enumerate(loading):
strain += strain_inc
alpha = np.sum(alpha_components)
yield_stress = sy_0 + Q * (1. - np.exp(-b * ep_eq)) - D * (1. - np.exp(-a * ep_eq))
trial_stress = stress + E * strain_inc
relative_stress = trial_stress - alpha
flow_dir = np.sign(relative_stress)
yield_condition = np.abs(relative_stress) - yield_stress
if yield_condition > tol:
is_converged = False
else:
is_converged = True
# Return mapping if plastic loading
ep_eq_init = ep_eq
alpha_init = alpha
consist_param = 0.
number_of_iterations = 0
while is_converged is False and number_of_iterations < maximum_iterations:
number_of_iterations += 1
# Isotropic hardening and isotropic modulus
yield_stress = sy_0 + Q * (1. - np.exp(-b * ep_eq)) - D * (1. - np.exp(-a * ep_eq))
iso_modulus = Q * b * np.exp(-b * ep_eq) - D * a * np.exp(-a * ep_eq)
# Kinematic hardening and kinematic modulus
alpha = 0.
kin_modulus = 0.
for i in range(0, n_backstresses):
e_k = np.exp(-gamma_k[i] * (ep_eq - ep_eq_init))
alpha += flow_dir * c_k[i] / gamma_k[i] + (alpha_components[i] - flow_dir * c_k[i] / gamma_k[i]) * e_k
kin_modulus += c_k[i] * e_k - flow_dir * gamma_k[i] * e_k * alpha_components[i]
delta_alpha = alpha - alpha_init
# Local Newton step
numerator = np.abs(relative_stress) - (consist_param * E + yield_stress + flow_dir * delta_alpha)
denominator = -(E + iso_modulus + kin_modulus)
consist_param = consist_param - numerator / denominator
ep_eq = ep_eq_init + consist_param
if np.abs(numerator) < tol:
is_converged = True
# Update the variables
stress = trial_stress - E * flow_dir * consist_param
for i in range(0, n_backstresses):
e_k = np.exp(-gamma_k[i] * (ep_eq - ep_eq_init))
alpha_components[i] = flow_dir * c_k[i] / gamma_k[i] \
+ (alpha_components[i] - flow_dir * c_k[i] / gamma_k[i]) * e_k
stress_track.append(stress)
strain_track.append(strain)
strain_inc_track.append(strain_inc)
iteration_track.append(number_of_iterations)
# Calculate the tangent modulus
if number_of_iterations > 0:
h_prime = 0.
for i in range(0, n_backstresses):
h_prime += c_k[i] - flow_dir * gamma_k[i] * alpha_components[i]
k_prime = Q * b * np.exp(-b * ep_eq) - D * a * np.exp(-a * ep_eq)
tangent_track.append(E * (k_prime + h_prime) / (E + k_prime + h_prime))
else:
# Elastic loading
tangent_track.append(E)
return np.append([0.], np.array(tangent_track))
| 38.755495 | 119 | 0.641455 | [
"MIT"
] | AlbanoCastroSousa/RESSPyLab | RESSPyLab/uvc_model.py | 14,107 | Python |
import wikipedia as wiki
from ..parsing import get_wiki_page_id, get_wiki_lines, get_wiki_sections
def get_wiki_references(url, outfile=None):
"""get_wiki_references.
Extracts references from predefined sections of wiki page
Uses `urlscan`, `refextract`, `doi`, `wikipedia`, and `re` (for ArXiv URLs)
:param url: URL of wiki article to scrape
:param outfile: File to write extracted references to
"""
def _check(l):
return (not l['doi'] or l['doi'] == l['refs'][-1]['doi']) \
and (not l['arxiv'] or l['arxiv'] == l['refs'][-1]['arxiv'])
page = wiki.page(get_wiki_page_id(url))
sections = get_wiki_sections(page.content)
lines = sum([get_wiki_lines(s, predicate=any) for s in sections.values()], [])
links = sum([wikiparse.parse(s).external_links for s in sections.values()], [])
summary = sum([
[
{
'raw': l,
'links': urlscan.parse_text_urls(l),
'refs': refextract.extract_references_from_string(l),
'doi': doi.find_doi_in_text(l),
'arxiv': m.group(1) if (m := arxiv_url_regex.matches(l)) is not None else None
} for l in get_wiki_lines(s, predicate=any)
] for s in sections.values()
])
failed = [ld for ld in summary if not _check(ld)]
if any(failed):
logger.warning('Consistency check failed for the following lines: {}'.format(failed))
return _serialize(summary, outfile)
if __name__ == "__main__":
import doctest
doctest.testmod()
| 41 | 94 | 0.622593 | [
"MIT"
] | antimike/citation-scraper | scraper/apis/wikipedia.py | 1,558 | Python |
from armulator.armv6.opcodes.abstract_opcodes.mov_register_thumb import MovRegisterThumb
from armulator.armv6.opcodes.opcode import Opcode
class MovRegisterThumbT2(MovRegisterThumb, Opcode):
def __init__(self, instruction, m, d):
Opcode.__init__(self, instruction)
MovRegisterThumb.__init__(self, True, m, d)
def is_pc_changing_opcode(self):
return self.d == 15
@staticmethod
def from_bitarray(instr, processor):
rd = instr[13:16]
rm = instr[10:13]
if processor.in_it_block():
print "unpredictable"
else:
return MovRegisterThumbT2(instr, **{"m": rm.uint, "d": rd.uint})
| 31.904762 | 88 | 0.674627 | [
"MIT"
] | AhmedMounir/armulator | armulator/armv6/opcodes/thumb_instruction_set/thumb_instruction_set_encoding_16_bit/thumb_shift_immediate_add_subtract_move_and_compare/mov_register_thumb_t2.py | 670 | Python |
class DefaultConfig(object):
DEBUG = False
JSONIFY_PRETTYPRINT_REGULAR = False
SQLALCHEMY_TRACK_MODIFICATIONS = False
API_PREFIX = '/api'
| 25.666667 | 42 | 0.74026 | [
"MIT"
] | sz-piotr/fioletowe-pomarancze | server/src/config.py | 154 | Python |
from concurrent.futures.process import ProcessPoolExecutor
import api.Config
import api.middleware
from api.Config import app
from api.routers import (feedback, hiscore, label, legacy, legacy_debug,
player, prediction, report, scraper)
app.include_router(hiscore.router)
app.include_router(player.router)
app.include_router(prediction.router)
app.include_router(feedback.router)
app.include_router(report.router)
app.include_router(legacy.router)
app.include_router(scraper.router)
app.include_router(label.router)
app.include_router(legacy_debug.router)
@app.get("/")
async def root():
return {"message": "Hello World"}
# @app.on_event("startup")
# async def startup_event():
# app.state.executor = ProcessPoolExecutor()
# @app.on_event("shutdown")
# async def on_shutdown():
# app.state.executor.shutdown()
| 25.878788 | 72 | 0.757611 | [
"BSD-2-Clause"
] | ThorntonMatthewD/Bot-Detector-Core-Files | api/app.py | 854 | Python |
# Copyright lowRISC contributors.
# Licensed under the Apache License, Version 2.0, see LICENSE for details.
# SPDX-License-Identifier: Apache-2.0
import os
import subprocess
OTBN_DIR = os.path.join(os.path.dirname(__file__), '../../..')
UTIL_DIR = os.path.join(OTBN_DIR, 'util')
SIM_DIR = os.path.join(os.path.dirname(__file__), '..')
def asm_and_link_one_file(asm_path: str, work_dir: str) -> str:
'''Assemble and link file at asm_path in work_dir.
Returns the path to the resulting ELF
'''
otbn_as = os.path.join(UTIL_DIR, 'otbn-as')
otbn_ld = os.path.join(UTIL_DIR, 'otbn-ld')
obj_path = os.path.join(work_dir, 'tst.o')
elf_path = os.path.join(work_dir, 'tst')
subprocess.run([otbn_as, '-o', obj_path, asm_path], check=True)
subprocess.run([otbn_ld, '-o', elf_path, obj_path], check=True)
return elf_path
| 30.607143 | 74 | 0.690782 | [
"Apache-2.0"
] | OneToughMonkey/opentitan | hw/ip/otbn/dv/otbnsim/test/testutil.py | 857 | Python |
import numpy as np
from .VariableUnitTest import VariableUnitTest
from gwlfe.MultiUse_Fxns.Discharge import UrbanQTotal
class TestUrbanQTotal(VariableUnitTest):
def test_UrbanQTotal(self):
z = self.z
np.testing.assert_array_almost_equal(
UrbanQTotal.UrbanQTotal_f(z.NYrs, z.DaysMonth, z.NRur, z.NUrb, z.Temp, z.InitSnow_0, z.Prec, z.Area,
z.CNI_0,
z.AntMoist_0, z.Grow_0, z.CNP_0, z.Imper, z.ISRR, z.ISRA),
UrbanQTotal.UrbanQTotal(z.NYrs, z.DaysMonth, z.NRur, z.NUrb, z.Temp, z.InitSnow_0, z.Prec, z.Area, z.CNI_0,
z.AntMoist_0, z.Grow_0, z.CNP_0, z.Imper, z.ISRR, z.ISRA), decimal=7)
| 43.823529 | 119 | 0.606711 | [
"Apache-2.0"
] | rajadain/gwlf-e | test/unittests/test_UrbanQTotal.py | 745 | Python |
#Find,Remove,Find
"""Return a tuple of the indices of the two smallest values in list L.
>>> items = [809, 834, 477, 478, 307, 122, 96, 102, 324, 476]
>>> find_two_smallest(items)
(6, 7)
>>> items == [809, 834, 477, 478, 307, 122, 96, 102, 324, 476]
True
"""
from typing import List, Tuple
def find_two_smallest(L:List[float]) -> Tuple[int, int]:
""" (see above) """
# Find the index of the minimum and remove that item
smallest = min(L)
min1 = L.index(smallest)
L.remove(smallest)
# Find the index of the new minimum item in the list
next_smallest = min(L)
min2 = L.index(next_smallest)
# Put smallest back into L
L.insert(min1, smallest)
# Fix min2 in case it was affected by the removal and reinsertion:
if min1 <= min2:
min2 +=1
return (min1, min2)
if __name__ == '__main__':
import doctest
doctest.testmod()
print(find_two_smallest([0, 1, 3, 2, 5, 6, 1]))
| 24.3 | 70 | 0.614198 | [
"MIT"
] | YordanIH/Intro_to_CS_w_Python | chapter12/examples/example02.py | 972 | Python |
from __future__ import absolute_import, print_function
import logging
import bokeh.server.tornado as tornado
from bokeh.application import Application
from bokeh.client import pull_session
from bokeh.server.views.static_handler import StaticHandler
from .utils import ManagedServerLoop, url
logging.basicConfig(level=logging.DEBUG)
def test_check_whitelist_rejects_port_mismatch():
assert False == tornado.check_whitelist("foo:100", ["foo:101", "foo:102"])
def test_check_whitelist_rejects_name_mismatch():
assert False == tornado.check_whitelist("foo:100", ["bar:100", "baz:100"])
def test_check_whitelist_accepts_name_port_match():
assert True == tornado.check_whitelist("foo:100", ["foo:100", "baz:100"])
def test_check_whitelist_accepts_implicit_port_80():
assert True == tornado.check_whitelist("foo", ["foo:80"])
def test_check_whitelist_accepts_all_on_star():
assert True == tornado.check_whitelist("192.168.0.1", ['*'])
assert True == tornado.check_whitelist("192.168.0.1:80", ['*'])
assert True == tornado.check_whitelist("192.168.0.1:5006", ['*'])
assert True == tornado.check_whitelist("192.168.0.1:80", ['*:80'])
assert False == tornado.check_whitelist("192.168.0.1:80", ['*:81'])
assert True == tornado.check_whitelist("192.168.0.1:5006", ['*:*'])
assert True == tornado.check_whitelist("192.168.0.1", ['192.168.0.*'])
assert True == tornado.check_whitelist("192.168.0.1:5006", ['192.168.0.*'])
assert False == tornado.check_whitelist("192.168.1.1", ['192.168.0.*'])
assert True == tornado.check_whitelist("foobarbaz", ['*'])
assert True == tornado.check_whitelist("192.168.0.1", ['192.168.0.*'])
assert False == tornado.check_whitelist("192.168.1.1", ['192.168.0.*'])
assert False == tornado.check_whitelist("192.168.0.1", ['192.168.0.*:5006'])
assert True == tornado.check_whitelist("192.168.0.1", ['192.168.0.*:80'])
assert True == tornado.check_whitelist("foobarbaz", ['*'])
assert True == tornado.check_whitelist("foobarbaz", ['*:*'])
assert True == tornado.check_whitelist("foobarbaz", ['*:80'])
assert False == tornado.check_whitelist("foobarbaz", ['*:5006'])
assert True == tornado.check_whitelist("foobarbaz:5006", ['*'])
assert True == tornado.check_whitelist("foobarbaz:5006", ['*:*'])
assert True == tornado.check_whitelist("foobarbaz:5006", ['*:5006'])
def test_default_resources():
application = Application()
with ManagedServerLoop(application) as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == ""
assert r.path_versioner == StaticHandler.append_version
with ManagedServerLoop(application, prefix="/foo/") as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == "/foo/"
assert r.path_versioner == StaticHandler.append_version
with ManagedServerLoop(application, prefix="foo/") as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == "/foo/"
assert r.path_versioner == StaticHandler.append_version
with ManagedServerLoop(application, prefix="foo") as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == "/foo/"
assert r.path_versioner == StaticHandler.append_version
with ManagedServerLoop(application, prefix="/foo") as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == "/foo/"
assert r.path_versioner == StaticHandler.append_version
with ManagedServerLoop(application, prefix="/foo/bar") as server:
r = server._tornado.resources()
assert r.mode == "server"
assert r.root_url == "/foo/bar/"
assert r.path_versioner == StaticHandler.append_version
def test_default_app_paths():
app = Application()
t = tornado.BokehTornado({}, "", [])
assert t.app_paths == set()
t = tornado.BokehTornado({"/": app}, "", [])
assert t.app_paths == { "/" }
t = tornado.BokehTornado({"/": app, "/foo": app}, "", [])
assert t.app_paths == { "/", "/foo"}
# tried to use capsys to test what's actually logged and it wasn't
# working, in the meantime at least this tests that log_stats
# doesn't crash in various scenarios
def test_log_stats():
application = Application()
with ManagedServerLoop(application) as server:
server._tornado.log_stats()
session1 = pull_session(session_id='session1',
url=url(server),
io_loop=server.io_loop)
session2 = pull_session(session_id='session2',
url=url(server),
io_loop=server.io_loop)
server._tornado.log_stats()
session1.close()
session2.close()
server._tornado.log_stats()
| 42.732759 | 80 | 0.65463 | [
"BSD-3-Clause"
] | Maluuba/bokeh | bokeh/server/tests/test_tornado.py | 4,957 | Python |
#!/usr/bin/env python
# Copyright 2015 Luminal, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import argparse
import csv
import json
import operator
import os
import os.path
import sys
import time
import re
import boto3
import botocore.exceptions
try:
from StringIO import StringIO
except ImportError:
from io import StringIO
try:
import yaml
NO_YAML = False
except ImportError:
NO_YAML = True
from base64 import b64encode, b64decode
from boto3.dynamodb.conditions import Attr
from Crypto.Cipher import AES
from Crypto.Hash import SHA256
from Crypto.Hash.HMAC import HMAC
from Crypto.Util import Counter
DEFAULT_REGION = "us-east-1"
PAD_LEN = 19 # number of digits in sys.maxint
WILDCARD_CHAR = "*"
class KmsError(Exception):
def __init__(self, value=""):
self.value = "KMS ERROR: " + value if value is not "" else "KMS ERROR"
def __str__(self):
return self.value
class IntegrityError(Exception):
def __init__(self, value=""):
self.value = "INTEGRITY ERROR: " + value if value is not "" else \
"INTEGRITY ERROR"
def __str__(self):
return self.value
class ItemNotFound(Exception):
pass
class KeyValueToDictionary(argparse.Action):
def __call__(self, parser, namespace, values, option_string=None):
setattr(namespace,
self.dest,
dict((x[0], x[1]) for x in values))
def printStdErr(s):
sys.stderr.write(str(s))
sys.stderr.write("\n")
def key_value_pair(string):
output = string.split('=')
if len(output) != 2:
msg = "%r is not the form of \"key=value\"" % string
raise argparse.ArgumentTypeError(msg)
return output
def expand_wildcard(string, secrets):
prog = re.compile('^' + string.replace(WILDCARD_CHAR, '.*') + '$')
output = []
for secret in secrets:
if prog.search(secret) is not None:
output.append(secret)
return output
def value_or_filename(string):
# argparse running on old version of python (<2.7) will pass an empty
# string to this function before it passes the actual value.
# If an empty string is passes in, just return an empty string
if string == "":
return ""
if string[0] == "@":
filename = string[1:]
try:
with open(os.path.expanduser(filename)) as f:
output = f.read()
except IOError as e:
raise argparse.ArgumentTypeError("Unable to read file %s" %
filename)
else:
output = string
return output
def csv_dump(dictionary):
csvfile = StringIO()
csvwriter = csv.writer(csvfile)
for key in dictionary:
csvwriter.writerow([key, dictionary[key]])
return csvfile.getvalue()
def paddedInt(i):
'''
return a string that contains `i`, left-padded with 0's up to PAD_LEN digits
'''
i_str = str(i)
pad = PAD_LEN - len(i_str)
return (pad * "0") + i_str
def getHighestVersion(name, region="us-east-1", table="credential-store"):
'''
Return the highest version of `name` in the table
'''
dynamodb = boto3.resource('dynamodb', region_name=region)
secrets = dynamodb.Table(table)
response = secrets.query(Limit=1,
ScanIndexForward=False,
ConsistentRead=True,
KeyConditionExpression=boto3.dynamodb.conditions.Key("name").eq(name),
ProjectionExpression="version")
if response["Count"] == 0:
return 0
return response["Items"][0]["version"]
def listSecrets(region="us-east-1", table="credential-store"):
'''
do a full-table scan of the credential-store,
and return the names and versions of every credential
'''
dynamodb = boto3.resource('dynamodb', region_name=region)
secrets = dynamodb.Table(table)
response = secrets.scan(ProjectionExpression="#N, version",
ExpressionAttributeNames={"#N": "name"})
return response["Items"]
def putSecret(name, secret, version, kms_key="alias/credstash",
region="us-east-1", table="credential-store", context=None):
'''
put a secret called `name` into the secret-store,
protected by the key kms_key
'''
if not context:
context = {}
kms = boto3.client('kms', region_name=region)
# generate a a 64 byte key.
# Half will be for data encryption, the other half for HMAC
try:
kms_response = kms.generate_data_key(KeyId=kms_key, EncryptionContext=context, NumberOfBytes=64)
except:
raise KmsError("Could not generate key using KMS key %s" % kms_key)
data_key = kms_response['Plaintext'][:32]
hmac_key = kms_response['Plaintext'][32:]
wrapped_key = kms_response['CiphertextBlob']
enc_ctr = Counter.new(128)
encryptor = AES.new(data_key, AES.MODE_CTR, counter=enc_ctr)
c_text = encryptor.encrypt(secret)
# compute an HMAC using the hmac key and the ciphertext
hmac = HMAC(hmac_key, msg=c_text, digestmod=SHA256)
b64hmac = hmac.hexdigest()
dynamodb = boto3.resource('dynamodb', region_name=region)
secrets = dynamodb.Table(table)
data = {}
data['name'] = name
data['version'] = version if version != "" else paddedInt(1)
data['key'] = b64encode(wrapped_key).decode('utf-8')
data['contents'] = b64encode(c_text).decode('utf-8')
data['hmac'] = b64hmac
return secrets.put_item(Item=data, ConditionExpression=Attr('name').not_exists())
def getAllSecrets(version="", region="us-east-1",
table="credential-store", context=None):
'''
fetch and decrypt all secrets
'''
output = {}
secrets = listSecrets(region, table)
for credential in set([x["name"] for x in secrets]):
try:
output[credential] = getSecret(credential,
version,
region,
table,
context)
except:
pass
return output
def getSecret(name, version="", region="us-east-1",
table="credential-store", context=None):
'''
fetch and decrypt the secret called `name`
'''
if not context:
context = {}
dynamodb = boto3.resource('dynamodb', region_name=region)
secrets = dynamodb.Table(table)
if version == "":
# do a consistent fetch of the credential with the highest version
response = secrets.query(Limit=1,
ScanIndexForward=False,
ConsistentRead=True,
KeyConditionExpression=boto3.dynamodb.conditions.Key("name").eq(name))
if response["Count"] == 0:
raise ItemNotFound("Item {'name': '%s'} couldn't be found." % name)
material = response["Items"][0]
else:
response = secrets.get_item(Key={"name": name, "version": version})
if "Item" not in response:
raise ItemNotFound("Item {'name': '%s', 'version': '%s'} couldn't be found." % (name, version))
material = response["Item"]
kms = boto3.client('kms', region_name=region)
# Check the HMAC before we decrypt to verify ciphertext integrity
try:
kms_response = kms.decrypt(CiphertextBlob=b64decode(material['key']), EncryptionContext=context)
except botocore.exceptions.ClientError as e:
if e.response["Error"]["Code"] == "InvalidCiphertextException":
if context is None:
msg = ("Could not decrypt hmac key with KMS. The credential may "
"require that an encryption context be provided to decrypt "
"it.")
else:
msg = ("Could not decrypt hmac key with KMS. The encryption "
"context provided may not match the one used when the "
"credential was stored.")
else:
msg = "Decryption error %s" % e
raise KmsError(msg)
except Exception as e:
raise KmsError("Decryption error %s" % e)
key = kms_response['Plaintext'][:32]
hmac_key = kms_response['Plaintext'][32:]
hmac = HMAC(hmac_key, msg=b64decode(material['contents']),
digestmod=SHA256)
if hmac.hexdigest() != material['hmac']:
raise IntegrityError("Computed HMAC on %s does not match stored HMAC"
% name)
dec_ctr = Counter.new(128)
decryptor = AES.new(key, AES.MODE_CTR, counter=dec_ctr)
plaintext = decryptor.decrypt(b64decode(material['contents'])).decode("utf-8")
return plaintext
def deleteSecrets(name, region="us-east-1", table="credential-store"):
dynamodb = boto3.resource('dynamodb', region_name=region)
secrets = dynamodb.Table(table)
response = secrets.scan(FilterExpression=boto3.dynamodb.conditions.Attr("name").eq(name),
ProjectionExpression="#N, version",
ExpressionAttributeNames={"#N": "name"})
for secret in response["Items"]:
print("Deleting %s -- version %s" % (secret["name"], secret["version"]))
secrets.delete_item(Key=secret)
def createDdbTable(region="us-east-1", table="credential-store"):
'''
create the secret store table in DDB in the specified region
'''
dynamodb = boto3.resource("dynamodb", region_name=region)
if table in (t.name for t in dynamodb.tables.all()):
print("Credential Store table already exists")
return
print("Creating table...")
response = dynamodb.create_table(
TableName=table,
KeySchema=[
{
"AttributeName": "name",
"KeyType": "HASH",
},
{
"AttributeName": "version",
"KeyType": "RANGE",
}
],
AttributeDefinitions=[
{
"AttributeName": "name",
"AttributeType": "S",
},
{
"AttributeName": "version",
"AttributeType": "S",
},
],
ProvisionedThroughput={
"ReadCapacityUnits": 1,
"WriteCapacityUnits": 1,
}
)
print("Waiting for table to be created...")
client = boto3.client("dynamodb", region_name=region)
client.get_waiter("table_exists").wait(TableName=table)
print("Table has been created. "
"Go read the README about how to create your KMS key")
def main():
parsers = {}
parsers['super'] = argparse.ArgumentParser(
description="A credential/secret storage system")
parsers['super'].add_argument("-r", "--region",
help="the AWS region in which to operate."
"If a region is not specified, credstash "
"will use the value of the "
"AWS_DEFAULT_REGION env variable, "
"or if that is not set, us-east-1")
parsers['super'].add_argument("-t", "--table", default="credential-store",
help="DynamoDB table to use for "
"credential storage")
subparsers = parsers['super'].add_subparsers(help='Try commands like '
'"{name} get -h" or "{name}'
'put --help" to get each'
'sub command\'s options'
.format(name=os.path.basename(
__file__)))
action = 'delete'
parsers[action] = subparsers.add_parser(action,
help='Delete a credential " \
"from the store')
parsers[action].add_argument("credential", type=str,
help="the name of the credential to delete")
parsers[action].set_defaults(action=action)
action = 'get'
parsers[action] = subparsers.add_parser(action, help="Get a credential "
"from the store")
parsers[action].add_argument("credential", type=str,
help="the name of the credential to get."
"Using the wildcard character '%s' will "
"search for credentials that match the "
"pattern" % WILDCARD_CHAR)
parsers[action].add_argument("context", type=key_value_pair,
action=KeyValueToDictionary, nargs='*',
help="encryption context key/value pairs "
"associated with the credential in the form "
"of \"key=value\"")
parsers[action].add_argument("-n", "--noline", action="store_true",
help="Don't append newline to returned "
"value (useful in scripts or with "
"binary files)")
parsers[action].add_argument("-v", "--version", default="",
help="Get a specific version of the "
"credential (defaults to the latest version)")
parsers[action].set_defaults(action=action)
action = 'getall'
parsers[action] = subparsers.add_parser(action,
help="Get all credentials from "
"the store")
parsers[action].add_argument("context", type=key_value_pair,
action=KeyValueToDictionary, nargs='*',
help="encryption context key/value pairs "
"associated with the credential in the form "
"of \"key=value\"")
parsers[action].add_argument("-v", "--version", default="",
help="Get a specific version of the "
"credential (defaults to the latest version)")
parsers[action].add_argument("-f", "--format", default="json",
choices=["json", "csv"] +
([] if NO_YAML else ["yaml"]),
help="Output format. json(default) " +
("" if NO_YAML else "yaml ") + "or csv.")
parsers[action].set_defaults(action=action)
action = 'list'
parsers[action] = subparsers.add_parser(action,
help="list credentials and "
"their versions")
parsers[action].set_defaults(action=action)
action = 'put'
parsers[action] = subparsers.add_parser(action,
help="Put a credential into "
"the store")
parsers[action].add_argument("credential", type=str,
help="the name of the credential to store")
parsers[action].add_argument("value", type=value_or_filename,
help="the value of the credential to store "
"or, if beginning with the \"@\" character, "
"the filename of the file containing "
"the value", default="")
parsers[action].add_argument("context", type=key_value_pair,
action=KeyValueToDictionary, nargs='*',
help="encryption context key/value pairs "
"associated with the credential in the form "
"of \"key=value\"")
parsers[action].add_argument("-k", "--key", default="alias/credstash",
help="the KMS key-id of the master key "
"to use. See the README for more "
"information. Defaults to alias/credstash")
parsers[action].add_argument("-v", "--version", default="",
help="Put a specific version of the "
"credential (update the credential; "
"defaults to version `1`).")
parsers[action].add_argument("-a", "--autoversion", action="store_true",
help="Automatically increment the version of "
"the credential to be stored. This option "
"causes the `-v` flag to be ignored. "
"(This option will fail if the currently stored "
"version is not numeric.)")
parsers[action].set_defaults(action=action)
action = 'setup'
parsers[action] = subparsers.add_parser(action,
help='setup the credential store')
parsers[action].set_defaults(action=action)
args = parsers['super'].parse_args()
region = os.getenv(
"AWS_DEFAULT_REGION", DEFAULT_REGION) if not args.region \
else args.region
if "action" in vars(args):
if args.action == "delete":
deleteSecrets(args.credential, region=region, table=args.table)
return
if args.action == "list":
credential_list = listSecrets(region=region, table=args.table)
if credential_list:
# print list of credential names and versions,
# sorted by name and then by version
max_len = max([len(x["name"]) for x in credential_list])
for cred in sorted(credential_list,
key=operator.itemgetter("name", "version")):
print("{0:{1}} -- version {2:>}".format(
cred["name"], max_len, cred["version"]))
else:
return
if args.action == "put":
if args.autoversion:
latestVersion = getHighestVersion(args.credential, region,
args.table)
try:
version = paddedInt(int(latestVersion) + 1)
except ValueError:
printStdErr("Can not autoincrement version. The current "
"version: %s is not an int" % latestVersion)
return
else:
version = args.version
try:
if putSecret(args.credential, args.value, version,
kms_key=args.key, region=region, table=args.table,
context=args.context):
print("{0} has been stored".format(args.credential))
except KmsError as e:
printStdErr(e)
except botocore.exceptions.ClientError as e:
if e.response["Error"]["Code"] == "ConditionalCheckFailedException":
latestVersion = getHighestVersion(args.credential, region,
args.table)
printStdErr("%s version %s is already in the credential store. "
"Use the -v flag to specify a new version" %
(args.credential, latestVersion))
return
if args.action == "get":
try:
if WILDCARD_CHAR in args.credential:
names = expand_wildcard(args.credential,
[x["name"]
for x
in listSecrets(region=region,
table=args.table)])
print(json.dumps(dict((name,
getSecret(name,
args.version,
region=region,
table=args.table,
context=args.context))
for name in names)))
else:
sys.stdout.write(getSecret(args.credential, args.version,
region=region, table=args.table,
context=args.context))
if not args.noline:
sys.stdout.write("\n")
except ItemNotFound as e:
printStdErr(e)
except KmsError as e:
printStdErr(e)
except IntegrityError as e:
printStdErr(e)
return
if args.action == "getall":
secrets = getAllSecrets(args.version,
region=region,
table=args.table,
context=args.context)
if args.format == "json":
output_func = json.dumps
output_args = {"sort_keys": True,
"indent": 4,
"separators": (',', ': ')}
elif not NO_YAML and args.format == "yaml":
output_func = yaml.dump
output_args = {"default_flow_style": False}
elif args.format == 'csv':
output_func = csv_dump
output_args = {}
print(output_func(secrets, **output_args))
return
if args.action == "setup":
createDdbTable(region=region, table=args.table)
return
else:
parsers['super'].print_help()
if __name__ == '__main__':
main()
| 40.668459 | 107 | 0.524567 | [
"Apache-2.0"
] | traveloka/credstash | credstash.py | 22,693 | Python |
host = "localhost"
port = 1111
max_users = 100
buffer_size = 1024
| 11.166667 | 18 | 0.716418 | [
"MIT"
] | iskrich/simplechat | params.py | 67 | Python |
"""
Django settings for lab01 project.
Generated by 'django-admin startproject' using Django 3.2.6.
For more information on this file, see
https://docs.djangoproject.com/en/3.2/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/3.2/ref/settings/
"""
from pathlib import Path
# Build paths inside the project like this: BASE_DIR / 'subdir'.
BASE_DIR = Path(__file__).resolve().parent.parent
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'django-insecure-7-8hv&pc-$$1)7eiiy2m#m^o6cx%oqqv9@z071ec0%218iwt0!'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'lab01.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'lab01.wsgi.application'
# Database
# https://docs.djangoproject.com/en/3.2/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': BASE_DIR / 'db.sqlite3',
}
}
# Password validation
# https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.2/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.2/howto/static-files/
STATIC_URL = '/static/'
# Default primary key field type
# https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field
DEFAULT_AUTO_FIELD = 'django.db.models.BigAutoField'
| 26.674603 | 92 | 0.675097 | [
"MIT"
] | car1os1/TECSUP-DAE-2021-2-B | lab01/lab01/settings.py | 3,361 | Python |
# Copyright 2018, Kay Hayen, mailto:[email protected]
#
# Python tests originally created or extracted from other peoples work. The
# parts were too small to be protected.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
def tupleUnpacking():
return (*a, b, *c)
def listUnpacking():
return [*a, b, *c]
def setUnpacking():
return {*a, b, *c}
def dictUnpacking():
return {"a" : 1, **d}
a = range(3)
b = 5
c = range(8,10)
d = {"a" : 2}
print("Tuple unpacked", tupleUnpacking())
print("List unpacked", listUnpacking())
print("Set unpacked", setUnpacking())
print("Dict unpacked", dictUnpacking())
non_iterable = 2.0
def tupleUnpackingError():
try:
return (*a,*non_iterable,*c)
except Exception as e:
return e
def listUnpackingError():
try:
return [*a,*non_iterable,*c]
except Exception as e:
return e
def setUnpackingError():
try:
return {*a,*non_iterable,*c}
except Exception as e:
return e
def dictUnpackingError():
try:
return {"a" : 1, **non_iterable}
except Exception as e:
return e
print("Tuple unpacked error:", tupleUnpackingError())
print("List unpacked error:", listUnpackingError())
print("Set unpacked error:", setUnpackingError())
print("Dict unpacked error:", dictUnpackingError())
| 25.561644 | 79 | 0.665595 | [
"Apache-2.0"
] | 4O4/Nuitka | tests/basics/Unpacking35.py | 1,866 | Python |
# Copyright (c) OpenMMLab. All rights reserved.
import torch.nn as nn
import torch.utils.checkpoint as cp
from mmcv.cnn import ConvModule
from mmcv.runner import BaseModule
from torch.nn.modules.batchnorm import _BatchNorm
from mmcls.models.utils import make_divisible
from ..builder import BACKBONES
from .base_backbone import BaseBackbone
class InvertedResidual(BaseModule):
"""InvertedResidual block for MobileNetV2.
Args:
in_channels (int): The input channels of the InvertedResidual block.
out_channels (int): The output channels of the InvertedResidual block.
stride (int): Stride of the middle (first) 3x3 convolution.
expand_ratio (int): adjusts number of channels of the hidden layer
in InvertedResidual by this amount.
conv_cfg (dict, optional): Config dict for convolution layer.
Default: None, which means using conv2d.
norm_cfg (dict): Config dict for normalization layer.
Default: dict(type='BN').
act_cfg (dict): Config dict for activation layer.
Default: dict(type='ReLU6').
with_cp (bool): Use checkpoint or not. Using checkpoint will save some
memory while slowing down the training speed. Default: False.
Returns:
Tensor: The output tensor
"""
def __init__(self,
in_channels,
out_channels,
stride,
expand_ratio,
conv_cfg=None,
norm_cfg=dict(type='BN'),
act_cfg=dict(type='ReLU'),
with_cp=False,
init_cfg=None):
super(InvertedResidual, self).__init__(init_cfg)
self.stride = stride
assert stride in [1, 2], f'stride must in [1, 2]. ' \
f'But received {stride}.'
self.with_cp = with_cp
self.use_res_connect = self.stride == 1 and in_channels == out_channels
hidden_dim = int(round(in_channels * expand_ratio))
layers = []
if expand_ratio != 1:
layers.append(
ConvModule(
in_channels=in_channels,
out_channels=hidden_dim,
kernel_size=1,
conv_cfg=conv_cfg,
norm_cfg=norm_cfg,
act_cfg=act_cfg))
layers.extend([
ConvModule(
in_channels=hidden_dim,
out_channels=hidden_dim,
kernel_size=3,
stride=stride,
padding=1,
groups=hidden_dim,
conv_cfg=conv_cfg,
norm_cfg=norm_cfg,
act_cfg=act_cfg),
ConvModule(
in_channels=hidden_dim,
out_channels=out_channels,
kernel_size=1,
conv_cfg=conv_cfg,
norm_cfg=norm_cfg,
act_cfg=None)
])
self.conv = nn.Sequential(*layers)
def forward(self, x):
def _inner_forward(x):
if self.use_res_connect:
return x + self.conv(x)
else:
return self.conv(x)
if self.with_cp and x.requires_grad:
out = cp.checkpoint(_inner_forward, x)
else:
out = _inner_forward(x)
return out
@BACKBONES.register_module()
class MobileNetV2(BaseBackbone):
"""MobileNetV2 backbone.
Args:
widen_factor (float): Width multiplier, multiply number of
channels in each layer by this amount. Default: 1.0.
out_indices (None or Sequence[int]): Output from which stages.
Default: (7, ).
frozen_stages (int): Stages to be frozen (all param fixed).
Default: -1, which means not freezing any parameters.
conv_cfg (dict, optional): Config dict for convolution layer.
Default: None, which means using conv2d.
norm_cfg (dict): Config dict for normalization layer.
Default: dict(type='BN').
act_cfg (dict): Config dict for activation layer.
Default: dict(type='ReLU6').
norm_eval (bool): Whether to set norm layers to eval mode, namely,
freeze running stats (mean and var). Note: Effect on Batch Norm
and its variants only. Default: False.
with_cp (bool): Use checkpoint or not. Using checkpoint will save some
memory while slowing down the training speed. Default: False.
"""
# Parameters to build layers. 4 parameters are needed to construct a
# layer, from left to right: expand_ratio, channel, num_blocks, stride.
arch_settings = [[1, 16, 1, 1], [6, 24, 2, 2], [6, 32, 3, 2],
[6, 64, 4, 2], [6, 96, 3, 1], [6, 160, 3, 2],
[6, 320, 1, 1]]
def __init__(self,
widen_factor=1.,
out_indices=(7, ),
frozen_stages=-1,
deep_stem=False,
conv_cfg=None,
norm_cfg=dict(type='BN'),
act_cfg=dict(type='ReLU'),
norm_eval=False,
with_cp=False,
init_cfg=[
dict(type='Kaiming', layer=['Conv2d']),
dict(
type='Constant',
val=1,
layer=['_BatchNorm', 'GroupNorm'])
]):
super(MobileNetV2, self).__init__(init_cfg)
self.widen_factor = widen_factor
self.out_indices = out_indices
for index in out_indices:
if index not in range(0, 8):
raise ValueError('the item in out_indices must in '
f'range(0, 8). But received {index}')
if frozen_stages not in range(-1, 8):
raise ValueError('frozen_stages must be in range(-1, 8). '
f'But received {frozen_stages}')
self.out_indices = out_indices
self.frozen_stages = frozen_stages
self.conv_cfg = conv_cfg
self.norm_cfg = norm_cfg
self.act_cfg = act_cfg
self.norm_eval = norm_eval
self.with_cp = with_cp
self.in_channels = make_divisible(32 * widen_factor, 8)
if deep_stem:
self.conv0 = ConvModule(in_channels=3, out_channels=16, kernel_size=3, stride=1, padding=1, conv_cfg=self.conv_cfg, norm_cfg=self.norm_cfg, act_cfg=self.act_cfg)
in_channels_ = 16
else:
in_channels_ = 3
self.conv0 = nn.Sequential()
self.conv1 = ConvModule(
in_channels=in_channels_,
out_channels=self.in_channels,
kernel_size=3,
stride=2,
padding=1,
conv_cfg=self.conv_cfg,
norm_cfg=self.norm_cfg,
act_cfg=self.act_cfg)
self.layers = []
for i, layer_cfg in enumerate(self.arch_settings):
expand_ratio, channel, num_blocks, stride = layer_cfg
out_channels = make_divisible(channel * widen_factor, 8)
inverted_res_layer = self.make_layer(
out_channels=out_channels,
num_blocks=num_blocks,
stride=stride,
expand_ratio=expand_ratio)
layer_name = f'layer{i + 1}'
self.add_module(layer_name, inverted_res_layer)
self.layers.append(layer_name)
if widen_factor > 1.0:
self.out_channel = int(1280 * widen_factor)
else:
self.out_channel = 1280
layer = ConvModule(
in_channels=self.in_channels,
out_channels=self.out_channel,
kernel_size=1,
stride=1,
padding=0,
conv_cfg=self.conv_cfg,
norm_cfg=self.norm_cfg,
act_cfg=self.act_cfg)
self.add_module('conv2', layer)
self.layers.append('conv2')
def make_layer(self, out_channels, num_blocks, stride, expand_ratio):
"""Stack InvertedResidual blocks to build a layer for MobileNetV2.
Args:
out_channels (int): out_channels of block.
num_blocks (int): number of blocks.
stride (int): stride of the first block. Default: 1
expand_ratio (int): Expand the number of channels of the
hidden layer in InvertedResidual by this ratio. Default: 6.
"""
layers = []
for i in range(num_blocks):
if i >= 1:
stride = 1
layers.append(
InvertedResidual(
self.in_channels,
out_channels,
stride,
expand_ratio=expand_ratio,
conv_cfg=self.conv_cfg,
norm_cfg=self.norm_cfg,
act_cfg=self.act_cfg,
with_cp=self.with_cp))
self.in_channels = out_channels
return nn.Sequential(*layers)
def forward(self, x):
x = self.conv0(x)
x = self.conv1(x)
outs = []
for i, layer_name in enumerate(self.layers):
layer = getattr(self, layer_name)
x = layer(x)
if i in self.out_indices:
outs.append(x)
return tuple(outs)
def _freeze_stages(self):
if self.frozen_stages >= 0:
for param in self.conv1.parameters():
param.requires_grad = False
for i in range(1, self.frozen_stages + 1):
layer = getattr(self, f'layer{i}')
layer.eval()
for param in layer.parameters():
param.requires_grad = False
def train(self, mode=True):
super(MobileNetV2, self).train(mode)
self._freeze_stages()
if mode and self.norm_eval:
for m in self.modules():
if isinstance(m, _BatchNorm):
m.eval()
| 36.639706 | 173 | 0.553783 | [
"Apache-2.0"
] | ChaseMonsterAway/mmclassification | mmcls/models/backbones/mobilenet_v2.py | 9,966 | Python |
import glob
import time
from os import path
from flask import Blueprint, jsonify, current_app, request, Response, json
from flask_login import login_required
from .. import pz_server_state
from ..services.power_actions_service import is_valid_power_action, execute_action
from ..services.server_options_service import read_config, save_config, prepared_config_to_view, formatted_config_lines
from ..services.server_status_service import get_server_status
from ..utils.resources_functions import server_resources
server_blueprint = Blueprint('server', __name__, url_prefix='/server')
@server_blueprint.route('/status')
@login_required
def status():
rcon_host = current_app.config['RCON_HOST']
rcon_password = current_app.config['RCON_PASSWORD']
server_state, players = get_server_status(rcon_host, rcon_password)
return jsonify(
server_state=server_state,
online_players=players,
server_resources=server_resources()
)
@server_blueprint.route('/power-actions', methods=['POST'])
@login_required
def power_actions():
request_data = request.get_json()
pz_user_home = current_app.config["PZ_USER_HOME"]
power_action = request_data.get("power_action", None)
if not is_valid_power_action(power_action):
return jsonify(error="Unknown action"), 400
if not execute_action(power_action, pz_user_home):
return '', 500
return jsonify(server_state=pz_server_state.state)
def get_config(pz_server_config):
config = read_config(pz_server_config)
return {
"WorkshopItems": config["WorkshopItems"],
"Mods": config["Mods"]
}
@server_blueprint.route('/options')
@login_required
def list_workshop_items():
export_config = get_config(current_app.config['PZ_SERVER_CONFIG'])
return jsonify(
WorkshopItems=prepared_config_to_view(export_config["WorkshopItems"]),
Mods=prepared_config_to_view(export_config["Mods"])
)
@server_blueprint.route('/options/export')
@login_required
def export_server_config():
export_config = get_config(current_app.config['PZ_SERVER_CONFIG'])
return current_app.response_class(
formatted_config_lines(export_config),
mimetype='text/event-stream',
headers={"Content-Disposition": "attachment;filename=server_config.ini"}
)
@server_blueprint.route('/options', methods=['POST'])
@login_required
def save_items():
request_data = request.get_json()
config = save_config(current_app.config['PZ_SERVER_CONFIG'], request_data)
export_config = {
"WorkshopItems": prepared_config_to_view(config["WorkshopItems"]),
"Mods": prepared_config_to_view(config["Mods"])
}
return jsonify(export_config)
@server_blueprint.route('/log')
@login_required
def listen_log():
def followLog(serverLogsDir):
logFilePattern = "*_DebugLog-server.txt"
logFiles = glob.glob(path.join(serverLogsDir, logFilePattern))
if not logFiles:
yield 'data: {}\n\n'.format(
json.dumps({"error": True, "errorMessage": "No log file found"})
)
return
logFiles.sort(reverse=True)
with open(logFiles[0]) as serverLogFile:
try:
while True:
line = serverLogFile.readline()
if not line:
continue
time.sleep(0.01)
yield 'data: {}\n\n'.format(
json.dumps({"log": line.strip()})
)
finally:
pass
serverLogsDir = current_app.config['PZ_SERVER_LOGS_DIR']
return Response(followLog(serverLogsDir), mimetype='text/event-stream')
| 29.8 | 119 | 0.68698 | [
"MIT"
] | emilio2hd/pz-panel | panel/routes/server.py | 3,725 | Python |
import pytest
from receptor.router import MeshRouter
test_networks = [
(
[
("a", "b", 1),
("a", "d", 1),
("a", "f", 1),
("b", "d", 1),
("b", "c", 1),
("c", "e", 1),
("c", "h", 1),
("c", "j", 1),
("e", "f", 1),
("e", "g", 1),
("e", "h", 1),
("f", "g", 1),
("g", "h", 1),
("h", "j", 1),
("h", "k", 1),
("j", "k", 1),
("j", "m", 1),
("l", "m", 1),
],
[("a", "f", "f"), ("a", "m", "b"), ("h", "d", "c")],
[("a", {"b", "d", "f"}), ("f", {"a", "e", "g"}), ("j", {"c", "h", "k", "m"})],
),
(
[("a", "b", 1), ("b", "c", 1), ("c", "d", 1), ("d", "e", 1), ("e", "f", 1)],
[("a", "f", "b"), ("c", "a", "b"), ("f", "c", "e")],
[("a", {"b"}), ("f", {"e"}), ("c", {"b", "d"})],
),
]
@pytest.mark.parametrize("edges, expected_next_hops, expected_neighbors", test_networks)
def test_next_hop(edges, expected_next_hops, expected_neighbors):
for node_id, remote, enh in expected_next_hops:
r = MeshRouter(node_id=node_id)
r.add_or_update_edges(edges)
assert r.next_hop(remote) == enh
@pytest.mark.parametrize("edges, expected_next_hops, expected_neighbors", test_networks)
def test_neighbors(edges, expected_next_hops, expected_neighbors):
r = MeshRouter(node_id=edges[0][0])
r.add_or_update_edges(edges)
for node_id, neighbors in expected_neighbors:
assert r.get_neighbors(node_id) == neighbors
| 31.745098 | 88 | 0.413218 | [
"Apache-2.0"
] | RedHatOfficial/receptor | test/unit/test_router.py | 1,619 | Python |
"""
多线程操作共享的全局变量是不安全的,多线程操作局部 只归某个线程私有,其他线程是不能访问的
"""
import threading
def do_sth(arg1, arg2, arg3):
local_var1 = arg1
local_var2 = arg2
local_var3 = arg3
fun1(local_var1, local_var2, local_var3)
fun2(local_var1, local_var2, local_var3)
fun3(local_var1, local_var2, local_var3)
def fun1(local_var1, local_var2, local_var3):
print('%s: %s -- %s -- %s' % (threading.current_thread().name, local_var1,
local_var2, local_var3))
def fun2(local_var1, local_var2, local_var3):
print('%s: %s -- %s -- %s' % (threading.current_thread().name, local_var1,
local_var2, local_var3))
def fun3(local_var1, local_var2, local_var3):
print('%s: %s -- %s -- %s' % (threading.current_thread().name, local_var1,
local_var2, local_var3))
t1 = threading.Thread(target=do_sth, args=('a', 'b', 'c'))
t2 = threading.Thread(target=do_sth, args=('d', 'e', 'f'))
t1.start()
t2.start()
| 27.189189 | 78 | 0.61332 | [
"Apache-2.0"
] | hemuke/python | 17_process_thread/46_why_need_ThreadLocal.py | 1,094 | Python |
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
# Copyright 2011-2019, Nigel Small
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from unittest import TestCase
from neotime import Date, Time, DateTime, Duration
from py2neo.data import Node
from py2neo.cypher import cypher_escape, cypher_repr
from py2neo.cypher.encoding import LabelSetView, PropertyDictView, PropertySelector
class LabelSetViewTestCase(TestCase):
def test_can_create_empty_view(self):
view = LabelSetView([])
self.assertEqual(repr(view), "")
def test_can_create_single_label_view(self):
view = LabelSetView(["A"])
self.assertEqual(repr(view), ":A")
def test_can_create_double_label_view(self):
view = LabelSetView(["A", "B"])
self.assertEqual(repr(view), ":A:B")
def test_can_select_existing_in_view(self):
view = LabelSetView(["A", "B"]).B
self.assertEqual(repr(view), ":B")
def test_can_select_non_existing_in_view(self):
view = LabelSetView(["A", "B"]).C
self.assertEqual(repr(view), "")
def test_can_chain_select(self):
view = LabelSetView(["A", "B", "C"]).B.C
self.assertEqual(repr(view), ":B:C")
def test_can_reselect_same(self):
view = LabelSetView(["A", "B", "C"]).B.B.C
self.assertEqual(repr(view), ":B:C")
def test_length(self):
view = LabelSetView(["A", "B", "C"])
self.assertEqual(len(view), 3)
def test_iterable(self):
view = LabelSetView(["A", "B", "C"])
self.assertSetEqual(set(view), {"A", "B", "C"})
def test_containment(self):
view = LabelSetView(["A", "B", "C"])
self.assertIn("A", view)
def test_non_containment(self):
view = LabelSetView(["A", "B", "C"])
self.assertNotIn("D", view)
class PropertyDictViewTestCase(TestCase):
def test_can_create_empty_view(self):
view = PropertyDictView({})
self.assertEqual(repr(view), "{}")
def test_can_create_single_property_view(self):
view = PropertyDictView({"A": 1})
self.assertEqual(repr(view), "{A: 1}")
def test_can_create_double_property_view(self):
view = PropertyDictView({"A": 1, "B": 2})
self.assertEqual(repr(view), "{A: 1, B: 2}")
def test_can_select_existing_in_view(self):
view = PropertyDictView({"A": 1, "B": 2}).B
self.assertEqual(repr(view), "{B: 2}")
def test_can_select_non_existing_in_view(self):
view = PropertyDictView({"A": 1, "B": 2}).C
self.assertEqual(repr(view), "{}")
def test_can_chain_select(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3}).B.C
self.assertEqual(repr(view), "{B: 2, C: 3}")
def test_can_reselect_same(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3}).B.B.C
self.assertEqual(repr(view), "{B: 2, C: 3}")
def test_length(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3})
self.assertEqual(len(view), 3)
def test_iterable(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3})
self.assertEqual(set(view), {"A", "B", "C"})
def test_containment(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3})
self.assertIn("A", view)
def test_non_containment(self):
view = PropertyDictView({"A": 1, "B": 2, "C": 3})
self.assertNotIn("D", view)
class PropertySelectorTestCase(TestCase):
def test_simple(self):
selector = PropertySelector({"A": 1, "B": 2, "C": 3})
self.assertEqual(selector.A, "1")
def test_non_existent(self):
selector = PropertySelector({"A": 1, "B": 2, "C": 3})
self.assertEqual(selector.D, "null")
class NodeReprTestCase(TestCase):
def test_empty(self):
a = Node()
r = cypher_repr(a)
self.assertEqual("({})", r)
def test_single_property(self):
a = Node(name="Alice")
r = cypher_repr(a)
self.assertEqual("({name: 'Alice'})", r)
def test_property_and_label(self):
a = Node("Person", name="Alice")
r = cypher_repr(a)
self.assertEqual("(:Person {name: 'Alice'})", r)
def test_date_property(self):
a = Node(d=Date(1970, 1, 1))
r = cypher_repr(a)
self.assertEqual("({d: date('1970-01-01')})", r)
def test_time_property(self):
a = Node(t=Time(12, 34, 56))
r = cypher_repr(a)
self.assertEqual("({t: time('12:34:56.000000000')})", r)
def test_datetime_property(self):
a = Node(dt=DateTime(1970, 1, 1, 12, 34, 56))
r = cypher_repr(a)
self.assertEqual("({dt: datetime('1970-01-01T12:34:56.000000000')})", r)
def test_duration_property(self):
a = Node(dur=Duration(days=3))
r = cypher_repr(a)
self.assertEqual("({dur: duration('P3D')})", r)
class CypherEscapeTestCase(TestCase):
def test_empty_string(self):
value = ""
with self.assertRaises(ValueError):
_ = cypher_escape(value)
def test_simple_string(self):
value = "foo"
escaped = "foo"
self.assertEqual(escaped, cypher_escape(value))
def test_string_with_space(self):
value = "foo bar"
escaped = "`foo bar`"
self.assertEqual(escaped, cypher_escape(value))
def test_string_with_backtick(self):
value = "foo `bar`"
escaped = "`foo ``bar```"
self.assertEqual(escaped, cypher_escape(value))
| 31.114583 | 83 | 0.61165 | [
"Apache-2.0"
] | CyberGRX/py2neo | test/unit/test_cypher_encoding.py | 5,974 | Python |
# Written by David Weber
# [email protected]
"""
In this short namespace I house a class that connects to PDB and downloads
file over PDB file transfer protocol.
"""
# ------------------------------------------------------------------------------
import gzip
from os import remove, getcwd, path # built in
# my pymol API built on Python2 - try both imports
try:
from urllib.request import urlretrieve, urlcleanup
except ImportError:
from urllib import urlretrieve, urlcleanup
ROOT = 'ftp://ftp.wwpdb.org/pub/pdb/data/structures/divided/pdb/{}/{}'
class PDBFile:
def __init__(self, code):
"""Initialize a PDBFile object with a pdb file of interest
Parameters
----------
code : the pdb code if interest
Any valid PDB code can be passed into PDBFile.
Examples
--------
>>> pdb_file = PDBFile('1rcy')
"""
self.code = code.lower()
def fetch_from_PDB(self):
"""
Connects to PDB FTP server, downloads a .gz file of interest,
decompresses the .gz file into .ent and then dumps a copy of
the pdb{code}.ent file into cwd.
Parameters
----------
None
Examples
--------
>>> inst = PDBFile('1rcy')
>>> path_to_file = inst.fetch_from_PDB()
>>> print(path_to_file)
"""
subdir = self.code[1:3]
infile = 'pdb{}.ent.gz'.format(self.code)
decompressed = infile.strip('.gz')
fullpath = ROOT.format(subdir, infile)
try:
urlcleanup()
urlretrieve(fullpath, infile)
except Exception:
return 'URLError'
else:
with gzip.open(infile, 'rb') as gz:
with open(decompressed, 'wb') as out:
out.writelines(gz)
remove(infile)
return path.join(getcwd(), decompressed)
def clear(self):
"""
Deletes file from current working directory after the file has
been processed by some algorithm.
Parameters
----------
None
Examples
--------
>>> inst = PDBFile('1rcy')
>>> path_to_file = inst.fetch_from_PDB()
>>> print(path_to_file) # process the file using some algorithm
>>> inst.clear()
"""
filename = 'pdb{}.ent'.format(self.code)
try:
remove(path.join(getcwd(), filename))
except FileNotFoundError:
print('Cannot delete file. Does not exist.')
| 28.222222 | 81 | 0.494273 | [
"MIT"
] | dsw7/BridgingInteractions | scalene-triangle/libs/PDB_filegetter.py | 2,794 | Python |
# -*- coding: utf-8 -*-
#Chucky_Bot
import LINETCR
from LINETCR.lib.curve.ttypes import *
from datetime import datetime
from bs4 import BeautifulSoup
from threading import Thread
from googletrans import Translator
from gtts import gTTS
import time,random,sys,json,codecs,threading,glob,urllib,urllib2,urllib3,re,ast,os,subprocess,requests,tempfile
cl = LINETCR.LINE()
#cl.login(qr=True)
cl.login(token='EsOvPPzeFykCVG8OoGf0.hE4TS1Hheb46PcdMzZKaaa.rzBOrFqSAApZownyv2qBJWU3PWWbf9/oE6G+sSVzUTo=')
cl.loginResult()
print "Azmi 1-Login Success\n"
ki = LINETCR.LINE()
#ki.login(qr=True)
ki.login(token='EsTdk3fyUSbT7LJVwoEd.rLylacrPH39WJb0UIwB8Nq.GYYzsgzj7aHd7mzCSluc3162Uqrry6Jjwf/bFuq9Etw=')
ki.loginResult()
print "Ki-Login Success\n"
kk = LINETCR.LINE()
#kk.login(qr=True)
kk.login(token='EsNKJDaP0J7Pt7syTOW9.GgPTp3/FisKkVX1rJHeroq.hUG0VDbWHz8R7o80xI0Pvme8dBb3dSsmCnat0PRX+JM=')
kk.loginResult()
print "Kk-Login Success\n"
#kc = LINETCR.LINE()
#kc.login(qr=True)
#kc.login(token='TOKEN_KAMU_DISINI_BEIB')
#kc.loginResult()
#print "Kc-Login Success\n"
#kr = LINETCR.LINE()
#kr.login(qr=True)
#kr.login(token='TOKEN_KAMU_DISINI_BEIB')
#kr.loginResult()
#print "Kr-Login Success\n"
#km = LINETCR.LINE()
#km.login(qr=True)
#km.login(token='TOKEN_KAMU_DISINI_BEIB')
#km.loginResult()
print "Km-Login Success\n\n=====[Sukses All Login]====="
reload(sys)
sys.setdefaultencoding('utf-8')
selfMessage ="""
╔═════════════════════════
║ ☆☞ S E L F ☜☆
╠═════════════════════════
╠➩〘Hi〙
╠➩〘Me〙
╠➩〘Mymid〙
╠➩〘Mid @〙
╠➩〘SearchID: (ID LINE)〙
╠➩〘Checkdate (DD/MM/YY)〙
╠➩〘Kalender〙
╠➩〘Steal contact〙
╠➩〘Pp @〙
╠➩〘Cover @〙
╠➩〘Auto like〙
╠➩〘Scbc Text〙
╠➩〘Cbc Text〙
╠➩〘Gbc Text〙
╠➩〘Getbio @〙
╠➩〘Getinfo @〙
╠➩〘Getname @〙
╠➩〘Getprofile @〙
╠➩〘Getcontact @〙
╠➩〘Getvid @〙
╠➩〘Friendlist〙
╠═════════════════════════
║ ☆☞ S E L F ☜☆
╚═════════════════════════
"""
botMessage ="""
╔═════════════════════════
║ ☆☞ B O T ☜☆
╠═════════════════════════
╠➩〘Absen〙
╠➩〘Respon〙
╠➩〘Runtime〙
╠➩〘Kapten copy @〙
╠➩〘TC1 copy @〙
╠➩〘TC2 copy @〙
╠➩〘TC3 copy @〙
╠➩〘TC4 copy @〙
╠➩〘Backup all〙
╠➩〘/bio Text〙
╠➩〘@bye (Usir Kapten)〙
╠➩〘Bye all (Usir Semua)〙
╠═════════════════════════
║ ☆☞ B O T ☜☆
╚═════════════════════════
"""
mediaMessage ="""
╔═════════════════════════
║ ☆☞ M E D I A ☜☆
╠═════════════════════════
╠➩〘Gift〙
╠➩〘Gift1 @ s/d Gift10 @〙
╠➩〘Giftbycontact〙
╠➩〘All gift〙
╠➩〘Gif gore〙
╠➩〘Google: (Text)〙
╠➩〘Playstore NamaApp〙
╠➩〘Fancytext: Text〙
╠➩〘/musik Judul-Penyanyi〙
╠➩〘/lirik Judul-Penyanyi〙
╠➩〘/musrik Judul-Penyanyi〙
╠➩〘/ig UrsnameInstagram〙
╠➩〘Checkig UrsnameInstagram〙
╠➩〘/apakah Text (Kerang Ajaib)〙
╠➩〘/kapan Text (Kerang Ajaib)〙
╠➩〘/hari Text (Kerang Ajaib)〙
╠➩〘/berapa Text (Kerang Ajaib)〙
╠➩〘/berapakah Text〙
╠➩〘Youtubelink: Judul Video〙
╠➩〘Youtubevideo: Judul Video〙
╠➩〘Youtubesearch: Judul Video〙
╠➩〘Image NamaGambar〙
╠➩〘Say-id Text〙
╠➩〘Say-en Text〙
╠➩〘Say-jp Text〙
╠➩〘Image NamaGambar〙
╠➩〘Tr-id Text (Translate En Ke ID〙
╠➩〘Tr-en Text (Translate ID Ke En〙
╠➩〘Tr-th Text (Translate ID Ke Th〙
╠➩〘Id@en Text (Translate ID Ke En〙
╠➩〘Id@th Text (Translate ID Ke TH〙
╠➩〘En@id Text (Translate En Ke ID〙
╠═════════════════════════
║ ☆☞ M E D I A ☜☆
╚═════════════════════════
"""
groupMessage ="""
╔═════════════════════════
║ ☆☞ G R O U P ☜☆
╠═════════════════════════
╠➩〘Welcome〙
╠➩〘Say welcome〙
╠➩〘Invite creator〙
╠➩〘Setview〙
╠➩〘Viewseen〙
╠➩〘Gn: (NamaGroup)〙
╠➩〘Tag all〙
╠➩〘Recover〙
╠➩〘Cancel〙
╠➩〘Cancelall〙
╠➩〘Gcreator〙
╠➩〘Ginfo〙
╠➩〘Gurl〙
╠➩〘List group〙
╠➩〘Pict group: (NamaGroup)〙
╠➩〘Spam: (Text)〙
╠➩〘Spam〙
╠➩〘Add all〙
╠➩〘Kick: (Mid)〙
╠➩〘Invite: (Mid)〙
╠➩〘Invite〙
╠➩〘Memlist〙
╠➩〘Getgroup image〙
╠➩〘Urlgroup Image〙
╠═════════════════════════
║ ☆☞ G R O U P ☜☆
╚═════════════════════════
"""
tjia="u71b6799e1c37868a871d442e67633182"
setMessage ="""
╔═════════════════════════
║ ☆☞ S E T ☜☆
╠═════════════════════════
╠➩〘Sambutan on/off〙
╠➩〘Url on/off〙
╠➩〘Alwaysread on/off〙
╠➩〘Sider on/off〙
╠➩〘Contact on/off〙
╠➩〘Simisimi on/off〙
╠═════════════════════════
║ ☆☞ S E T ☜☆
╚═════════════════════════
"""
creatorMessage ="""
╔═════════════════════════
║ ☆☞ C R E A T O R ☜☆
╠═════════════════════════
╠➩〘Admin add @〙
╠➩〘Admin remove @〙
╠➩〘/cnkapten〙
╠➩〘/cntc1〙
╠➩〘/cntc2〙
╠➩〘/cntc3〙
╠➩〘/cntc4〙
╠➩〘Crash〙
╠➩〘Kickall〙
╠➩〘Bc: (Text)〙
╠➩〘Nk: @〙
╠➩〘Ulti @〙
╠➩〘Join group: (NamaGroup〙
╠➩〘Leave group: (NamaGroup〙
╠➩〘Leave all group〙
╠➩〘Bot restart〙
╠➩〘Turn off〙
╠═════════════════════════
║ ☆☞ C R E A T O R ☜☆
╚═════════════════════════
"""
adminMessage ="""
╔═════════════════════════
║ ☆☞ A D M I N ☜☆
╠═════════════════════════
╠➩〘Admin list〙
╠➩〘Ban〙
╠➩〘Unban〙
╠➩〘Ban @〙
╠➩〘Unban @〙
╠➩〘Ban list〙
╠➩〘Clear ban〙
╠➩〘Kill〙
╠➩〘Kick @〙
╠➩〘Set member: (Jumblah)〙
╠➩〘Ban group: (NamaGroup〙
╠➩〘Del ban: (NamaGroup〙
╠➩〘List ban〙
╠➩〘Kill ban〙
╠➩〘Glist〙
╠➩〘Glistmid〙
╠➩〘Details group: (Gid)〙
╠➩〘Cancel invite: (Gid)〙
╠➩〘Invitemeto: (Gid)〙
╠➩〘Kapten acc invite〙
╠➩〘TC1 acc invite〙
╠➩〘TC2 acc invite〙
╠➩〘TC3 acc invite〙
╠➩〘TC4 acc invite〙
╠➩〘Removechat〙
╠➩〘Join on/off〙
╠➩〘Joincancel on/off〙
╠➩〘Respon on/off〙
╠➩〘Responkick on/off〙
╠➩〘Leave on/off〙
╠➩〘All join / (TC1/2/3/4 Join)〙
╠═════════════════════════
║ ☆☞ A D M I N ☜☆
╚═════════════════════════
"""
helpMessage ="""
╔═════════════════════════
║ ☆☞ H E L P ☜☆
╠═════════════════════════
╠➩〘Help protect〙
╠➩〘Help self〙
╠➩〘Help bot〙
╠➩〘Help group〙
╠➩〘Help set〙
╠➩〘Help media〙
╠➩〘Help admin〙
╠➩〘Help creator〙
╠➩〘Owner〙
╠➩〘Pap owner〙
╠➩〘Admin〙
╠➩〘Speed〙
╠➩〘Speed test〙
╠➩〘Status〙
╠═════════════════════════
║ ☆☞ H E L P ☜☆
╚═════════════════════════
"""
protectMessage ="""
╔═════════════════════════
║ ☆☞ P R O T E C T ☜☆
╠═════════════════════════
╠➩〘Allprotect on/off〙
╠➩〘Autocancel on/off〙
╠➩〘Qr on/off〙
╠➩〘Autokick on/off〙
╠➩〘Ghost on/off〙
╠➩〘Invitepro on/off〙
╠═════════════════════════
║ ☆☞ P R O T E C T ☜☆
╚═════════════════════════
"""
KAC=[cl,ki,kk]
mid = cl.getProfile().mid
Amid = ki.getProfile().mid
Bmid = kk.getProfile().mid
Bots=[mid,Amid,Bmid]
Creator=["u71b6799e1c37868a871d442e67633182"]
admin=["u71b6799e1c37868a871d442e67633182"]
contact = cl.getProfile()
backup1 = cl.getProfile()
backup1.displayName = contact.displayName
backup1.statusMessage = contact.statusMessage
backup1.pictureStatus = contact.pictureStatus
contact = ki.getProfile()
backup2 = ki.getProfile()
backup2.displayName = contact.displayName
backup2.statusMessage = contact.statusMessage
backup2.pictureStatus = contact.pictureStatus
contact = kk.getProfile()
backup3 = kk.getProfile()
backup3.displayName = contact.displayName
backup3.statusMessage = contact.statusMessage
backup3.pictureStatus = contact.pictureStatus
responsename = cl.getProfile().displayName
responsename2 = ki.getProfile().displayName
responsename3 = kk.getProfile().displayName
wait = {
"LeaveRoom":True,
"AutoJoin":False,
"AutoJoinCancel":True,
"memberscancel":0,
"Members":1,
"AutoCancel":{},
"AutoCancelon":False,
"joinkick":False,
"AutoKick":{},
"AutoKickon":False,
'pap':{},
'invite':{},
'steal':{},
'gift':{},
'likeOn':{},
'Leave':{},
'detectMention':True,
'kickMention':False,
'timeline':True,
"Timeline":True,
"comment1":"Kenapa Kak?",
"comment2":"Wkwkwk \(○^ω^○)/",
"comment3":"Lucu Banget!!! ヘ(^_^)ヘ",
"comment4":"Nice Kak (^_^)",
"comment5":"Bot Auto Like ©By : Azmi\nContact Me : 👉 line.me/ti/p/~a_ulul15",
"commentOn":True,
"commentBlack":{},
"message":"Thx For Add Me (^_^)\nInvite Me To Your Group ヘ(^_^)ヘ",
"blacklist":{},
"wblacklist":False,
"dblacklist":False,
"Qr":{},
"Qron":False,
"Contact":False,
"Sambutan":True,
"Ghost":False,
"inviteprotect":False,
"alwaysRead":False,
"Sider":{},
"Simi":{},
"lang":"JP",
"BlGroup":{}
}
settings = {
"simiSimi":{}
}
cctv = {
"cyduk":{},
"point":{},
"sidermem":{}
}
wait2 = {
"readPoint":{},
"readMember":{},
"setTime":{},
"ROM":{}
}
setTime = {}
setTime = wait2['setTime']
mulai = time.time()
def download_page(url):
version = (3,0)
cur_version = sys.version_info
if cur_version >= version:
import urllib,request
try:
headers = {}
headers['User-Agent'] = "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2228.0 Safari/537.36"
req = urllib,request.Request(url, headers = headers)
resp = urllib,request.urlopen(req)
respData = str(resp.read())
return respData
except Exception as e:
print(str(e))
else:
import urllib2
try:
headers = {}
headers['User-Agent'] = "Mozilla/5.0 (X11; Linux i686) AppleWebKit/537.17 (KHTML, like Gecko) Chrome/24.0.1312.27 Safari/537.17"
req = urllib2.Request(url, headers = headers)
response = urllib2.urlopen(req)
page = response.read()
return page
except:
return"Page Not found"
def _images_get_next_item(s):
start_line = s.find('rg_di')
if start_line == -1:
end_quote = 0
link = "no_links"
return link, end_quote
else:
start_line = s.find('"class="rg_meta"')
start_content = s.find('"ou"',start_line+90)
end_content = s.find(',"ow"',start_content-90)
content_raw = str(s[start_content+6:end_content-1])
return content_raw, end_content
def _images_get_all_items(page):
items = []
while True:
item, end_content = _images_get_next_item(page)
if item == "no_links":
break
else:
items.append(item)
time.sleep(0.1)
page = page[end_content:]
return items
def waktu(secs):
mins, secs = divmod(secs,60)
hours, mins = divmod(mins,60)
return '%02d Jam %02d Menit %02d Detik' % (hours, mins, secs)
def cms(string, commands):# /XXX, >XXX, ;XXX, ^XXX, %XXX, $XXX...
tex = ["+","@","/",">",";","^","%","$","^","サテラ:","サテラ:","サテラ:","サテラ:"]
for texX in tex:
for command in commands:
if string ==command:
return True
return False
def upload_tempimage(client):
'''
Upload a picture of a kitten. We don't ship one, so get creative!
'''
config = {
'album': album,
'name': 'bot auto upload',
'title': 'bot auto upload',
'description': 'bot auto upload'
}
print("Uploading image... ")
image = client.upload_from_path(image_path, config=config, anon=False)
print("Done")
print()
return image
def sendAudio(self, to_, path):
M = Message()
M.text = None
M.to = to_
M.contentMetadata = None
M.contentPreview = None
M.contentType = 3
M_id = self._client.sendMessage(0,M).id
files = {
'file': open(path, 'rb'),
}
def sendMessage(to, text, contentMetadata={}, contentType=0):
mes = Message()
mes.to, mes.from_ = to, profile.mid
mes.text = text
mes.contentType, mes.contentMetadata = contentType, contentMetadata
if to not in messageReq:
messageReq[to] = -1
messageReq[to] += 1
def sendImage(self, to_, path):
M = Message(to=to_, text=None, contentType = 1)
M.contentMetadata = None
M.contentPreview = None
M2 = self._client.sendMessage(0,M)
M_id = M2.id
files = {
'file': open(path, 'rb'),
}
params = {
'name': 'media',
'oid': M_id,
'size': len(open(path, 'rb').read()),
'type': 'image',
'ver': '1.0',
}
data = {
'params': json.dumps(params)
}
r = self.post_content('https://obs-sg.line-apps.com/talk/m/upload.nhn', data=data, files=files)
if r.status_code != 201:
raise Exception('Upload image failure.')
return True
def sendImageWithURL(self, to_, url):
path = '%s/pythonLine-%i.data' % (tempfile.gettempdir(), randint(0, 9))
r = requests.get(url, stream=True)
if r.status_code == 200:
with open(path, 'w') as f:
shutil.copyfileobj(r.raw, f)
else:
raise Exception('Download image failure.')
try:
self.sendImage(to_, path)
except:
try:
self.sendImage(to_, path)
except Exception as e:
raise e
def sendAudio(self, to_, path):
M = Message()
M.text = None
M.to = to_
M.contentMetadata = None
M.contentPreview = None
M.contentType = 3
M_id = self._client.sendMessage(0,M).id
files = {
'file': open(path, 'rb'),
}
params = {
'name': 'media',
'oid': M_id,
'size': len(open(path, 'rb').read()),
'type': 'audio',
'ver': '1.0',
}
data = {
'params': json.dumps(params)
}
r = self.post_content('https://os.line.naver.jp/talk/m/upload.nhn', data=data, files=files)
if r.status_code != 201:
raise Exception('Upload audio failure.')
return True
def sendAudioWithURL(self, to_, url):
path = self.downloadFileWithURL(url)
try:
self.sendAudio(to_, path)
except Exception as e:
raise Exception(e)
def sendAudioWithUrl(self, to_, url):
path = '%s/pythonLine-%1.data' % (tempfile.gettempdir(), randint(0, 9))
r = requests.get(url, stream=True, verify=False)
if r.status_code == 200:
with open(path, 'w') as f:
shutil.copyfileobj(r.raw, f)
else:
raise Exception('Download audio failure.')
try:
self.sendAudio(to_, path)
except Exception as e:
raise e
def downloadFileWithURL(self, fileUrl):
saveAs = '%s/pythonLine-%i.data' % (tempfile.gettempdir(), randint(0, 9))
r = self.get_content(fileUrl)
if r.status_code == 200:
with open(saveAs, 'wb') as f:
shutil.copyfileobj(r.raw, f)
return saveAs
else:
raise Exception('Download file failure.')
def restart_program():
python = sys.executable
os.execl(python, python, * sys.argv)
def bot(op):
try:
if op.type == 0:
return
if op.type == 5:
if wait["autoAdd"] == True:
cl.findAndAddContactsByMid(op.param1)
if(wait["message"]in[""," ","\n",None]):
pass
else:
cl.sendText(op.param1,str(wait["message"]))
if op.type == 55:
try:
group_id = op.param1
user_id=op.param2
subprocess.Popen('echo "'+ user_id+'|'+str(op.createdTime)+'" >> dataSeen/%s.txt' % group_id, shell=True, stdout=subprocess.PIPE, )
except Exception as e:
print e
if op.type == 55:
try:
if cctv['cyduk'][op.param1]==True:
if op.param1 in cctv['point']:
Name = cl.getContact(op.param2).displayName
Name = ki.getContact(op.param2).displayName
Name = kk.getContact(op.param2).displayName
Name = kc.getContact(op.param2).displayName
Name = kr.getContact(op.param2).displayName
if Name in cctv['sidermem'][op.param1]:
pass
else:
cctv['sidermem'][op.param1] += "\n• " + Name
if " " in Name:
nick = Name.split(' ')
if len(nick) == 2:
random.choice(KAC).sendText(op.param1, "Haii " + "☞ " + nick[0] + " ☜" + "\nNgintip Aja Niih. . .\nChat Kek Idiih (-__-) ")
else:
random.choice(KAC).sendText(op.param1, "Haii " + "☞ " + nick[1] + " ☜" + "\nBetah Banget Jadi Penonton. . .\nChat Napa (-__-) ")
else:
random.choice(KAC).sendText(op.param1, "Haii " + "☞ " + Name + " ☜" + "\nNgapain Kak Ngintip Aja???\nSini Gabung Chat... ")
else:
pass
else:
pass
except:
pass
else:
pass
if op.type == 22:
cl.leaveRoom(op.param1)
if op.type == 21:
cl.leaveRoom(op.param1)
if op.type == 13:
print op.param3
if op.param3 in mid:
if op.param2 in Creator:
cl.acceptGroupInvitation(op.param1)
if op.param3 in Amid:
if op.param2 in Creator:
ki.acceptGroupInvitation(op.param1)
if op.param3 in Bmid:
if op.param2 in Creator:
kk.acceptGroupInvitation(op.param1)
if op.param3 in Cmid:
if op.param2 in Creator:
kc.acceptGroupInvitation(op.param1)
if op.param3 in Dmid:
if op.param2 in Creator:
kr.acceptGroupInvitation(op.param1)
if op.param3 in mid:
if op.param2 in Amid:
cl.acceptGroupInvitation(op.param1)
if op.param3 in mid:
if op.param2 in Bmid:
cl.acceptGroupInvitation(op.param1)
if op.param3 in mid:
if op.param2 in Cmid:
cl.acceptGroupInvitation(op.param1)
if op.param3 in Amid:
if op.param2 in mid:
ki.acceptGroupInvitation(op.param1)
if op.param3 in Amid:
if op.param2 in Bmid:
ki.acceptGroupInvitation(op.param1)
if op.param3 in Amid:
if op.param2 in Cmid:
ki.acceptGroupInvitation(op.param1)
if op.param3 in Bmid:
if op.param2 in mid:
kk.acceptGroupInvitation(op.param1)
if op.param3 in Bmid:
if op.param2 in Amid:
kk.acceptGroupInvitation(op.param1)
if op.param3 in Bmid:
if op.param2 in Cmid:
kk.acceptGroupInvitation(op.param1)
if op.param3 in Cmid:
if op.param2 in mid:
kc.acceptGroupInvitation(op.param1)
if op.param3 in Cmid:
if op.param2 in Amid:
kc.acceptGroupInvitation(op.param1)
if op.param3 in Cmid:
if op.param2 in Cmid:
kc.acceptGroupInvitation(op.param1)
if op.param3 in Dmid:
if op.param2 in mid:
kr.acceptGroupInvitation(op.param1)
if op.param3 in Dmid:
if op.param2 in Amid:
kr.acceptGroupInvitation(op.param1)
if op.param3 in Dmid:
if op.param2 in Bmid:
kr.acceptGroupInvitation(op.param1)
if mid in op.param3:
if wait["AutoJoinCancel"] == True:
G = cl.getGroup(op.param1)
if len(G.members) <= wait["memberscancel"]:
cl.acceptGroupInvitation(op.param1)
cl.sendText(op.param1,"Maaf " + cl.getContact(op.param2).displayName + "\nMember Kurang Dari 30 Orang\nUntuk Info, Silahkan Chat Owner Kami!")
c = Message(to=op.param1, from_=None, text=None, contentType=13)
c.contentMetadata={'mid':tjia}
cl.sendMessage(c)
cl.leaveGroup(op.param1)
else:
cl.acceptGroupInvitation(op.param1)
G = cl.getGroup(op.param1)
G.preventJoinByTicket = False
cl.updateGroup(G)
Ti = cl.reissueGroupTicket(op.param1)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G.preventJoinByTicket = True
cl.updateGroup(G)
cl.sendText(op.param1,"☆Ketik ☞Help☜ Untuk Bantuan☆\n☆Harap Gunakan Dengan Bijak ^_^ ☆")
if mid in op.param3:
if wait["AutoJoin"] == True:
G = cl.getGroup(op.param1)
if len(G.members) <= wait["Members"]:
cl.rejectGroupInvitation(op.param1)
else:
cl.acceptGroupInvitation(op.param1)
G = cl.getGroup(op.param1)
G.preventJoinByTicket = False
cl.updateGroup(G)
Ti = cl.reissueGroupTicket(op.param1)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G.preventJoinByTicket = True
cl.updateGroup(G)
cl.sendText(op.param1,"☆Ketik ☞Help☜ Untuk Bantuan☆\n☆Harap Gunakan Dengan Bijak ^_^ ☆")
else:
if wait["AutoCancel"][op.param1] == True:
if op.param3 in admin:
pass
else:
cl.cancelGroupInvitation(op.param1, [op.param3])
else:
if op.param3 in wait["blacklist"]:
cl.cancelGroupInvitation(op.param1, [op.param3])
cl.sendText(op.param1, "Blacklist Detected")
else:
pass
if op.type == 19:
if wait["AutoKick"][op.param1] == True:
try:
if op.param3 in Creator:
if op.param3 in admin:
if op.param3 in Bots:
pass
if op.param2 in Creator:
if op.param2 in admin:
if op.param2 in Bots:
pass
else:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
if op.param2 in wait["blacklist"]:
pass
else:
random.choice(KAC).inviteIntoGroup(op.param1,[op.param3])
except:
try:
if op.param2 not in Creator:
if op.param2 not in admin:
if op.param2 not in Bots:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
if op.param2 in wait["blacklist"]:
pass
else:
random.choice(KAC).inviteIntoGroup(op.param1,[op.param3])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Creator:
if op.param2 in admin:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Creator:
if op.param2 in admin:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
else:
pass
if mid in op.param3:
if op.param2 in Creator:
if op.param2 in Bots:
pass
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
G = ki.getGroup(op.param1)
G.preventJoinByTicket = False
ki.updateGroup(G)
Ti = ki.reissueGroupTicket(op.param1)
cl.acceptGroupInvitationByTicket(op.param1,Ti)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
X = cl.getGroup(op.param1)
X.preventJoinByTicket = True
cl.updateGroup(X)
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if Amid in op.param3:
if op.param2 in Bots:
pass
try:
kk.kickoutFromGroup(op.param1,[op.param2])
kc.kickoutFromGroup(op.param1,[op.param2])
except:
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
X = kk.getGroup(op.param1)
X.preventJoinByTicket = False
cl.updateGroup(X)
Ti = kk.reissueGroupTicket(op.param1)
cl.acceptGroupInvitationByTicket(op.param1,Ti)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G = ki.getGroup(op.param1)
G.preventJoinByTicket = True
ki.updateGroup(G)
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if Bmid in op.param3:
if op.param2 in Bots:
pass
try:
kc.kickoutFromGroup(op.param1,[op.param2])
kk.kickoutFromGroup(op.param1,[op.param2])
except:
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
X = kc.getGroup(op.param1)
X.preventJoinByTicket = False
kc.updateGroup(X)
Ti = kc.reissueGroupTicket(op.param1)
cl.acceptGroupInvitationByTicket(op.param1,Ti)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G = kk.getGroup(op.param1)
G.preventJoinByTicket = True
kk.updateGroup(G)
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if Cmid in op.param3:
if op.param2 in Bots:
pass
try:
cl.kickoutFromGroup(op.param1,[op.param2])
kk.kickoutFromGroup(op.param1,[op.param2])
except:
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
X = cl.getGroup(op.param1)
X.preventJoinByTicket = False
cl.updateGroup(X)
Ti = cl.reissueGroupTicket(op.param1)
cl.acceptGroupInvitationByTicket(op.param1,Ti)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G = kc.getGroup(op.param1)
G.preventJoinByTicket = True
kc.updateGroup(G)
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if Dmid in op.param3:
if op.param2 in Bots:
pass
try:
cl.kickoutFromGroup(op.param1,[op.param2])
kk.kickoutFromGroup(op.param1,[op.param2])
except:
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
X = cl.getGroup(op.param1)
X.preventJoinByTicket = False
cl.updateGroup(X)
Ti = cl.reissueGroupTicket(op.param1)
cl.acceptGroupInvitationByTicket(op.param1,Ti)
ki.acceptGroupInvitationByTicket(op.param1,Ti)
kk.acceptGroupInvitationByTicket(op.param1,Ti)
kc.acceptGroupInvitationByTicket(op.param1,Ti)
kr.acceptGroupInvitationByTicket(op.param1,Ti)
G = kc.getGroup(op.param1)
G.preventJoinByTicket = True
kc.updateGroup(G)
if op.param2 in wait["blacklist"]:
pass
else:
if op.param2 in Bots:
pass
else:
wait["blacklist"][op.param2] = True
if Creator in op.param3:
if admin in op.param3:
if op.param2 in Bots:
pass
try:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
except:
try:
if op.param2 not in Bots:
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
if op.param2 in wait["blacklist"]:
pass
else:
random.choice(KAC).inviteIntoGroup(op.param1,[op.param3])
except:
print ("client Kick regulation or Because it does not exist in the group\ngid=["+op.param1+"]\nmid=["+op.param2+"]")
if op.param2 in wait["blacklist"]:
pass
if op.param2 in wait["whitelist"]:
pass
else:
wait["blacklist"][op.param2] = True
random.choice(KAC).inviteIntoGroup(op.param1,[op.param3])
if op.param2 in wait["blacklist"]:
pass
if op.param2 in wait["whitelist"]:
pass
else:
wait["blacklist"][op.param2] = True
if op.type == 11:
if wait["Qr"][op.param1] == True:
if op.param2 not in Bots:
if op.param2 not in admin:
G = random.choice(KAC).getGroup(op.param1)
G.preventJoinByTicket = True
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
random.choice(KAC).updateGroup(G)
if op.type == 17:
if wait["Sambutan"] == True:
if op.param2 in admin:
return
ginfo = cl.getGroup(op.param1)
contact = cl.getContact(op.param2)
image = "http://dl.profile.line-cdn.net/" + contact.pictureStatus
c = Message(to=op.param1, from_=None, text=None, contentType=13)
c.contentMetadata={'mid':op.param2}
cl.sendMessage(c)
cl.sendText(op.param1,"Hallo " + cl.getContact(op.param2).displayName + "\nWelcome To ☞ " + str(ginfo.name) + " ☜" + "\nBudayakan Cek Note\nDan Semoga Betah Disini ^_^")
cl.sendImageWithURL(op.param1,image)
print "MEMBER JOIN TO GROUP"
if op.type == 17:
if wait["joinkick"] == True:
if op.param2 in admin:
if op.param2 in Bots:
return
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
print "MEMBER JOIN KICK TO GROUP"
if op.type == 15:
if wait["Sambutan"] == True:
if op.param2 in admin:
return
cl.sendText(op.param1,"Good Bye " + cl.getContact(op.param2).displayName + "\nSee You Next Time . . . (p′︵‵。) 🤗")
random.choice(KAC).inviteIntoGroup(op.param1,[op.param2])
print "MEMBER HAS LEFT THE GROUP"
if op.type == 13:
if op.param2 not in Creator:
if op.param2 not in admin:
if op.param2 not in Bots:
if op.param2 in Creator:
if op.param2 in admin:
if op.param2 in Bots:
pass
elif wait["inviteprotect"] == True:
wait ["blacklist"][op.param2] = True
cl.cancelGroupInvitation(op.param1,[op.param3])
random.choice(KAC).kickoutFromGroup(op.param1,[op.param2])
if op.type == 19:
if wait["Ghost"] == True:
if op.param2 in admin:
if op.param2 in Bots:
pass
else:
try:
G = cl.getGroup(op.param1)
G.preventJoinByTicket = False
cl.updateGroup(G)
Ticket = cl.reissueGroupTicket(op.param1)
km.acceptGroupInvitationByTicket(op.param1,Ticket)
time.sleep(0.01)
km.kickoutFromGroup(op.param1,[op.param2])
c = Message(to=op.param1, from_=None, text=None, contentType=13)
c.contentMetadata={'mid':op.param2}
km.sendMessage(c)
km.leaveGroup(op.param1)
G.preventJoinByTicket = True
cl.updateGroup(G)
wait["blacklist"][op.param2] = True
except:
G = cl.getGroup(op.param1)
G.preventJoinByTicket = False
cl.updateGroup(G)
Ticket = cl.reissueGroupTicket(op.param1)
km.acceptGroupInvitationByTicket(op.param1,Ticket)
time.sleep(0.01)
km.kickoutFromGroup(op.param1,[op.param2])
c = Message(to=op.param1, from_=None, text=None, contentType=13)
c.contentMetadata={'mid':op.param2}
km.sendMessage(c)
km.leaveGroup(op.param1)
G.preventJoinByTicket = True
cl.updateGroup(G)
wait["blacklist"][op.param2] = True
if op.type == 26:
msg = op.message
if wait["alwaysRead"] == True:
if msg.toType == 0:
cl.sendChatChecked(msg.from_,msg.id)
else:
cl.sendChatChecked(msg.to,msg.id)
if msg.contentType == 16:
if wait['likeOn'] == True:
url = msg.contentMetadata["postEndUrl"]
cl.like(url[25:58], url[66:], likeType=1005)
ki.like(url[25:58], url[66:], likeType=1002)
kk.like(url[25:58], url[66:], likeType=1004)
kc.like(url[25:58], url[66:], likeType=1003)
kr.like(url[25:58], url[66:], likeType=1001)
cl.comment(url[25:58], url[66:], wait["comment1"])
ki.comment(url[25:58], url[66:], wait["comment2"])
kk.comment(url[25:58], url[66:], wait["comment3"])
kc.comment(url[25:58], url[66:], wait["comment4"])
kr.comment(url[25:58], url[66:], wait["comment5"])
cl.sendText(msg.to,"Like Success")
wait['likeOn'] = False
if op.type == 26:
msg = op.message
if msg.to in settings["simiSimi"]:
if settings["simiSimi"][msg.to] == True:
if msg.text is not None:
text = msg.text
r = requests.get("http://api.ntcorp.us/chatbot/v1/?text=" + text.replace(" ","+") + "&key=beta1.nt")
data = r.text
data = json.loads(data)
if data['status'] == 200:
if data['result']['result'] == 100:
cl.sendText(msg.to,data['result']['response'].encode('utf-8'))
if 'MENTION' in msg.contentMetadata.keys() != None:
if wait["kickMention"] == True:
contact = cl.getContact(msg.from_)
cName = contact.displayName
balas = ["Aku Bilang Jangan Ngetag Lagi " + cName + "\nAku Kick Kamu! Sorry, Byee!!!"]
ret_ = random.choice(balas)
name = re.findall(r'@(\w+)', msg.text)
mention = ast.literal_eval(msg.contentMetadata['MENTION'])
mentionees = mention['MENTIONEES']
for mention in mentionees:
if mention['M'] in admin:
cl.sendText(msg.to,ret_)
random.choice(KAC).kickoutFromGroup(msg.to,[msg.from_])
break
if mention['M'] in Bots:
cl.sendText(msg.to,ret_)
random.choice(KAC).kickoutFromGroup(msg.to,[msg.from_])
break
if 'MENTION' in msg.contentMetadata.keys() != None:
if wait["detectMention"] == True:
contact = cl.getContact(msg.from_)
cName = contact.displayName
balas = ["Sekali lagi nge tag gw sumpahin jomblo seumur hidup!","Dont Tag!! Lagi Sibuk",cName + " Ngapain Ngetag?",cName + " Nggak Usah Tag-Tag! Kalo Penting Langsung Pc Aja","Tag Mulu Lo Anjirr!","Dia Lagi Off", cName + " Kenapa Tag? Kangen?","Dia Lagi Tidur\nJangan Di Tag " + cName, "Jangan Suka Tag Gua " + cName, "Kamu Siapa " + cName + "?", "Ada Perlu Apa " + cName + "?","Woii " + cName + " Jangan Ngetag, Riibut!"]
ret_ = random.choice(balas)
name = re.findall(r'@(\w+)', msg.text)
mention = ast.literal_eval(msg.contentMetadata['MENTION'])
mentionees = mention['MENTIONEES']
for mention in mentionees:
if mention['M'] in admin:
cl.sendText(msg.to,ret_)
break
if mention['M'] in Bots:
cl.sendText(msg.to,ret_)
break
if msg.contentType == 13:
if wait["wblacklist"] == True:
if msg.contentMetadata["mid"] not in admin:
if msg.contentMetadata["mid"] in wait["blacklist"]:
random.choice(KAC).sendText(msg.to,"Sudah")
wait["wblacklist"] = False
else:
wait["blacklist"][msg.contentMetadata["mid"]] = True
wait["wblacklist"] = False
random.choice(KAC).sendText(msg.to,"Ditambahkan")
else:
cl.sendText(msg.to,"Admin Detected~")
elif wait["dblacklist"] == True:
if msg.contentMetadata["mid"] in wait["blacklist"]:
del wait["blacklist"][msg.contentMetadata["mid"]]
random.choice(KAC).sendText(msg.to,"Terhapus")
wait["dblacklist"] = False
else:
wait["dblacklist"] = False
random.choice(KAC).sendText(msg.to,"Tidak Ada Black List")
elif wait["Contact"] == True:
msg.contentType = 0
cl.sendText(msg.to,msg.contentMetadata["mid"])
if 'displayName' in msg.contentMetadata:
contact = cl.getContact(msg.contentMetadata["mid"])
try:
cu = cl.channel.getCover(msg.contentMetadata["mid"])
except:
cu = ""
cl.sendText(msg.to,"Nama:\n" + msg.contentMetadata["displayName"] + "\n\nMid:\n" + msg.contentMetadata["mid"] + "\n\nStatus:\n" + contact.statusMessage + "\n\nPhoto Profile:\nhttp://dl.profile.line-cdn.net/" + contact.pictureStatus + "\n\nPhoto Cover:\n" + str(cu))
else:
contact = cl.getContact(msg.contentMetadata["mid"])
try:
cu = cl.channel.getCover(msg.contentMetadata["mid"])
except:
cu = ""
cl.sendText(msg.to,"Nama:\n" + msg.contentMetadata["displayName"] + "\n\nMid:\n" + msg.contentMetadata["mid"] + "\n\nStatus:\n" + contact.statusMessage + "\n\nPhoto Profile:\nhttp://dl.profile.line-cdn.net/" + contact.pictureStatus + "\n\nPhoto Cover:\n" + str(cu))
elif msg.text == "Ginfo":
if msg.toType == 2:
ginfo = cl.getGroup(msg.to)
try:
gCreator = ginfo.creator.displayName
except:
gCreator = "Error"
if wait["lang"] == "JP":
if ginfo.invitee is None:
sinvitee = "0"
else:
sinvitee = str(len(ginfo.invitee))
if ginfo.preventJoinByTicket == True:
u = "close"
else:
u = "open"
cl.sendText(msg.to,"[Group name]\n" + str(ginfo.name) + "\n\n[Gid]\n" + msg.to + "\n\n[Group creator]\n" + gCreator + "\n\n[Profile status]\nhttp://dl.profile.line.naver.jp/" + ginfo.pictureStatus + "\n\nMembers:" + str(len(ginfo.members)) + "members\nPending:" + sinvitee + "people\nURL:" + u + "it is inside")
else:
cl.sendText(msg.to,"[group name]\n" + str(ginfo.name) + "\n[gid]\n" + msg.to + "\n[group creator]\n" + gCreator + "\n[profile status]\nhttp://dl.profile.line.naver.jp/" + ginfo.pictureStatus)
else:
if wait["lang"] == "JP":
cl.sendText(msg.to,"Can not be used outside the group")
else:
cl.sendText(msg.to,"Not for use less than group")
elif msg.text is None:
return
elif msg.text in ["Creator","Owner"]:
msg.contentType = 13
msg.contentMetadata = {'mid': tjia}
cl.sendMessage(msg)
cl.sendText(msg.to,"Itu tukang tikungnya(^_^)")
elif msg.text in ["Admin","admin"]:
msg.contentType = 13
admin1 = "u71b6799e1c37868a871d442e67633182"
admin2 = "u46560b002469877f708c1d2e8966fc9d"
admin3 = "u1dee2db35847101e3aa420e667390000"
msg.contentMetadata = {'mid': tjia}
random.choice(KAC).sendMessage(msg)
msg.contentMetadata = {'mid': admin1}
random.choice(KAC).sendMessage(msg)
msg.contentMetadata = {'mid': admin2}
random.choice(KAC).sendMessage(msg)
msg.contentMetadata = {'mid': admin3}
random.choice(KAC).sendMessage(msg)
random.choice(KAC).sendText(msg.to,"Itu Admin Kami (^_^)")
elif "Admin add @" in msg.text:
if msg.from_ in Creator:
print "[Command]Admin add executing"
_name = msg.text.replace("Admin add @","")
_nametarget = _name.rstrip(' ')
gs = cl.getGroup(msg.to)
gs = ki.getGroup(msg.to)
gs = kk.getGroup(msg.to)
gs = kc.getGroup(msg.to)
gs = kr.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
random.choice(KAC).sendText(msg.to,"Contact Tidak Di Temukan")
else:
for target in targets:
try:
admin.append(target)
cl.sendText(msg.to,"Admin Chucky Ditambahkan")
except:
pass
print "[Command]Admin add executed"
else:
cl.sendText(msg.to,"Command Denied.")
cl.sendText(msg.to,"Creator Permission Required.")
elif "Admin remove @" in msg.text:
if msg.from_ in Creator:
print "[Command]Admin Remove Executing"
_name = msg.text.replace("Admin remove @","")
_nametarget = _name.rstrip(' ')
gs = cl.getGroup(msg.to)
gs = ki.getGroup(msg.to)
gs = kk.getGroup(msg.to)
gs = kc.getGroup(msg.to)
gs = kr.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
random.choice(KAC).sendText(msg.to,"Contact Tidak Di Temukan")
else:
for target in targets:
try:
admin.remove(target)
cl.sendText(msg.to,"Admin Chucky Dihapus")
except:
pass
print "[Command]Admin remove executed"
else:
cl.sendText(msg.to,"Command Denied.")
cl.sendText(msg.to,"Creator Permission Required.")
elif msg.text in ["Admin list","admin list","List admin"]:
if admin == []:
cl.sendText(msg.to,"The Admin List Is Empty")
else:
cl.sendText(msg.to,"Tunggu...")
mc = "╔═════════════════════════\n║ ☆☞ ADMIN CHUCKY ☜☆\n╠═════════════════════════\n"
for mi_d in admin:
mc += "╠••> " +cl.getContact(mi_d).displayName + "\n"
cl.sendText(msg.to,mc + "╚═════════════════════════")
print "[Command]Admin List executed"
elif msg.text in ["Group creator","Gcreator","gcreator"]:
ginfo = cl.getGroup(msg.to)
gCreator = ginfo.creator.mid
msg.contentType = 13
msg.contentMetadata = {'mid': gCreator}
cl.sendMessage(msg)
cl.sendText(msg.to,"Itu Yang Buat Grup Ini")
elif msg.contentType == 16:
if wait["Timeline"] == True:
msg.contentType = 0
msg.text = "post URL\n" + msg.contentMetadata["postEndUrl"]
random.choice(KAC).sendText(msg.to,msg.text)
if msg.contentType == 13:
if wait["steal"] == True:
_name = msg.contentMetadata["displayName"]
copy = msg.contentMetadata["mid"]
groups = cl.getGroup(msg.to)
pending = groups.invitee
targets = []
for s in groups.members:
if _name in s.displayName:
print "[Target] Stealed"
break
else:
targets.append(copy)
if targets == []:
pass
else:
for target in targets:
try:
cl.findAndAddContactsByMid(target)
contact = cl.getContact(target)
cu = cl.channel.getCover(target)
path = str(cu)
image = "http://dl.profile.line-cdn.net/" + contact.pictureStatus
cl.sendText(msg.to,"Nama :\n" + contact.displayName + "\n\nMid :\n" + msg.contentMetadata["mid"] + "\n\nBio :\n" + contact.statusMessage)
cl.sendText(msg.to,"Profile Picture " + contact.displayName)
cl.sendImageWithURL(msg.to,image)
cl.sendText(msg.to,"Cover " + contact.displayName)
cl.sendImageWithURL(msg.to,path)
wait["steal"] = False
break
except:
pass
if msg.contentType == 13:
if wait["gift"] == True:
_name = msg.contentMetadata["displayName"]
copy = msg.contentMetadata["mid"]
groups = cl.getGroup(msg.to)
pending = groups.invitee
targets = []
for s in groups.members:
if _name in s.displayName:
print "[Target] Gift"
break
else:
targets.append(copy)
if targets == []:
pass
else:
for target in targets:
try:
cl.sendText(msg.to,"Gift Sudah Terkirim!")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '1',
'STKPKGID': '1296261'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
wait['gift'] = False
break
except:
msg.contentMetadata = {'mid': target}
wait["gift"] = False
break
if msg.contentType == 13:
if wait['invite'] == True:
_name = msg.contentMetadata["displayName"]
invite = msg.contentMetadata["mid"]
groups = cl.getGroup(msg.to)
groups = ki.getGroup(msg.to)
groups = kk.getGroup(msg.to)
groups = kc.getGroup(msg.to)
groups = kr.getGroup(msg.to)
pending = groups.invitee
targets = []
for s in groups.members:
if _name in s.displayName:
random.choice(KAC).sendText(msg.to, _name + " Berada DiGrup Ini")
else:
targets.append(invite)
if targets == []:
pass
else:
for target in targets:
try:
cl.findAndAddContactsByMid(target)
ki.findAndAddContactsByMid(target)
kk.findAndAddContactsByMid(target)
kc.findAndAddContactsByMid(target)
kr.findAndAddContactsByMid(target)
random.choice(KAC).inviteIntoGroup(msg.to,[target])
random.choice(KAC).sendText(msg.to,"Invite " + _name)
wait['invite'] = False
break
except:
random.choice(KAC).sendText(msg.to,"Limit Invite")
wait['invite'] = False
break
elif msg.text in ["Key creator","help creator","Help creator"]:
cl.sendText(msg.to,creatorMessage)
elif msg.text in ["Key group","help group","Help group"]:
cl.sendText(msg.to,groupMessage)
elif msg.text in ["Key","help","Help"]:
cl.sendText(msg.to,helpMessage)
elif msg.text in ["Key self","help self","Help self"]:
cl.sendText(msg.to,selfMessage)
elif msg.text in ["Key bot","help bot","Help bot"]:
cl.sendText(msg.to,botMessage)
elif msg.text in ["Key set","help set","Help set"]:
cl.sendText(msg.to,setMessage)
elif msg.text in ["Key media","help media","Help media"]:
cl.sendText(msg.to,mediaMessage)
elif msg.text in ["Key admin","help admin","Help admin"]:
cl.sendText(msg.to,adminMessage)
elif msg.text in ["Key protect","help protect","Help protect"]:
cl.sendText(msg.to,protectMessage)
elif msg.text in ["List group"]:
gid = cl.getGroupIdsJoined()
h = ""
jml = 0
for i in gid:
gn = cl.getGroup(i).name
h += "♦【%s】\n" % (gn)
jml += 1
cl.sendText(msg.to,"=======[List Group]=======\n"+ h +"\nTotal Group: "+str(jml))
elif "Ban group: " in msg.text:
grp = msg.text.replace("Ban group: ","")
gid = cl.getGroupIdsJoined()
if msg.from_ in admin:
for i in gid:
h = cl.getGroup(i).name
if h == grp:
wait["BlGroup"][i]=True
cl.sendText(msg.to, "Success Ban Group : "+grp)
else:
pass
else:
cl.sendText(msg.to, "Only Admin")
elif msg.text in ["List ban","List ban group"]:
if msg.from_ in admin:
if wait["BlGroup"] == {}:
random.choice(KAC).sendText(msg.to,"Tidak Ada")
else:
mc = ""
for gid in wait["BlGroup"]:
mc += "-> " +cl.getGroup(gid).name + "\n"
random.choice(KAC).sendText(msg.to,"===[Ban Group]===\n"+mc)
else:
cl.sendText(msg.to, "Khusus Admin")
elif msg.text in ["Del ban: "]:
if msg.from_ in admin:
ng = msg.text.replace("Del ban: ","")
for gid in wait["BlGroup"]:
if cl.getGroup(gid).name == ng:
del wait["BlGroup"][gid]
cl.sendText(msg.to, "Success del ban "+ng)
else:
pass
else:
cl.sendText(msg.to, "Only Admin")
elif "Join group: " in msg.text:
ng = msg.text.replace("Join group: ","")
gid = cl.getGroupIdsJoined()
gid = ki.getGroupIdsJoined()
gid = kk.getGroupIdsJoined()
gid = kc.getGroupIdsJoined()
gid = kr.getGroupIdsJoined()
try:
if msg.from_ in Creator:
for i in gid:
h = cl.getGroup(i).name
h = ki.getGroup(i).name
h = kk.getGroup(i).name
h = kc.getGroup(i).name
h = kr.getGroup(i).name
if h == ng:
random.choice(KAC).inviteIntoGroup(i,[Creator])
cl.sendText(msg.to,"Success Join To ["+ h +"] Group")
else:
pass
else:
cl.sendText(msg.to,"Only Admin")
except Exception as e:
cl.sendText(msg.to, str(e))
elif "Leave group: " in msg.text:
ng = msg.text.replace("Leave group: ","")
gid = cl.getGroupIdsJoined()
if msg.from_ in Creator:
for i in gid:
h = cl.getGroup(i).name
if h == ng:
cl.sendText(i,"Bot Di Paksa Keluar Oleh Owner!")
cl.leaveGroup(i)
ki.leaveGroup(i)
kk.leaveGroup(i)
kc.leaveGroup(i)
kr.leaveGroup(i)
cl.sendText(msg.to,"Success Left ["+ h +"] group")
else:
pass
else:
cl.sendText(msg.to,"Only Admin")
elif "Leave all group" == msg.text:
gid = cl.getGroupIdsJoined()
if msg.from_ in Creator:
for i in gid:
cl.sendText(i,"Bot Di Paksa Keluar Oleh Owner!")
cl.leaveGroup(i)
ki.leaveGroup(i)
kk.leaveGroup(i)
kc.leaveGroup(i)
kr.leaveGroup(i)
cl.sendText(msg.to,"Success Leave All Group")
else:
cl.sendText(msg.to,"Only Admin")
elif "Pict group: " in msg.text:
saya = msg.text.replace('Pict group: ','')
gid = cl.getGroupIdsJoined()
for i in gid:
h = cl.getGroup(i).name
gna = cl.getGroup(i)
if h == saya:
cl.sendImageWithURL(msg.to,"http://dl.profile.line.naver.jp/"+ gna.pictureStatus)
elif msg.text in ["cancelall","Cancelall"]:
if msg.toType == 2:
X = cl.getGroup(msg.to)
if X.invitee is not None:
gInviMids = [contact.mid for contact in X.invitee]
cl.cancelGroupInvitation(msg.to, gInviMids)
else:
cl.sendText(msg.to,"Tidak Ada Yang Pending")
else:
cl.sendText(msg.to,"Tidak Bisa Digunakan Diluar Group")
elif msg.text in ["Ourl","Url on"]:
if msg.toType == 2:
X = cl.getGroup(msg.to)
X.preventJoinByTicket = False
cl.updateGroup(X)
cl.sendText(msg.to,"Url Sudah Aktif")
else:
cl.sendText(msg.to,"Can not be used outside the group")
elif msg.text in ["Curl","Url off"]:
if msg.toType == 2:
X = cl.getGroup(msg.to)
X.preventJoinByTicket = True
cl.updateGroup(X)
cl.sendText(msg.to,"Url Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Can not be used outside the group")
elif msg.text in ["Join on","Autojoin on"]:
if msg.from_ in admin:
wait["AutoJoin"] = True
wait["AutoJoinCancel"] = False
cl.sendText(msg.to,"Auto Join Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Join off","Autojoin off"]:
if msg.from_ in admin:
wait["AutoJoin"] = False
cl.sendText(msg.to,"Auto Join Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Joincancel on","Autojoincancel on"]:
if msg.from_ in admin:
wait["AutoJoinCancel"] = True
wait["AutoJoin"] = False
cl.sendText(msg.to,"Auto Join Cancel Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Joincancel off","Autojoincancel off"]:
if msg.from_ in admin:
wait["AutoJoinCancel"] = False
cl.sendText(msg.to,"Auto Join Cancel Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Respon on"]:
if msg.from_ in admin:
wait["detectMention"] = True
wait["kickMention"] = False
cl.sendText(msg.to,"Auto Respon Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Respon off"]:
if msg.from_ in admin:
wait["detectMention"] = False
cl.sendText(msg.to,"Auto Respon Sudah Off")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Responkick on"]:
if msg.from_ in admin:
wait["kickMention"] = True
wait["detectMention"] = False
cl.sendText(msg.to,"Auto Respon Kick Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Responkick off"]:
if msg.from_ in admin:
wait["kickMention"] = False
cl.sendText(msg.to,"Auto Respon Kick Sudah Off")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Leave on"]:
if msg.from_ in admin:
wait["Leave"] = True
cl.sendText(msg.to,"Leave Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Autocancel on"]:
if msg.from_ in admin:
wait["AutoCancel"][msg.to] = True
wait["AutoCancelon"] = True
cl.sendText(msg.to,"Auto Cancel Sudah Aktif")
print wait["AutoCancel"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Autocancel off"]:
if msg.from_ in admin:
wait["AutoCancel"][msg.to] = False
wait["AutoCancelon"] = False
cl.sendText(msg.to,"Auto Cancel Sudah Di Nonaktifkan")
print wait["AutoCancel"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Joinkick on"]:
if msg.from_ in admin:
wait["joinkick"] = True
wait["Sambutan"] = False
cl.sendText(msg.to,"Join Kick Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Joinkick off"]:
if msg.from_ in admin:
wait["joinkick"] = False
cl.sendText(msg.to,"Join Kick Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Invitepro on","Inviteprotect on"]:
if msg.from_ in admin:
wait["inviteprotect"] = True
cl.sendText(msg.to,"Invite Protect Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Invitepro off","Inviteprotect off"]:
if msg.from_ in admin:
wait["inviteprotect"] = False
cl.sendText(msg.to,"Invite Protect Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Only Admin")
elif "Qr on" in msg.text:
if msg.from_ in admin:
wait["Qr"][msg.to] = True
wait["Qron"] = True
cl.sendText(msg.to,"QR Protect Sudah Aktif")
print wait["Qr"]
else:
cl.sendText(msg.to,"Only Admin")
elif "Qr off" in msg.text:
if msg.from_ in admin:
wait["Qr"][msg.to] = False
wait["Qron"] = False
cl.sendText(msg.to,"Qr Protect Sudah Di Nonaktifkan")
print wait["Qr"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Autokick on"]:
if msg.from_ in admin:
wait["AutoKick"][msg.to] = True
wait["AutoKickon"] = True
cl.sendText(msg.to,"Auto Kick Sudah Aktif")
print wait["AutoKick"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Autokick off"]:
if msg.from_ in admin:
wait["AutoKick"][msg.to] = False
wait["AutoKickon"] = False
cl.sendText(msg.to,"Auto Kick Sudah Di Nonaktifkan")
print wait["AutoKick"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Ghost on"]:
if msg.from_ in admin:
wait["Ghost"] = True
cl.sendText(msg.to,"Ghost Sudah Aktif")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Ghost off"]:
if msg.from_ in admin:
wait["Ghost"] = False
cl.sendText(msg.to,"Ghost Sudah Di Nonaktifkan")
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Allprotect on"]:
if msg.from_ in admin:
wait["AutoCancel"][msg.to] = True
wait["AutoCancelon"] = True
wait["inviteprotect"] = True
wait["joinkick"] = True
wait["AutoKick"][msg.to] = True
wait["AutoKickon"] = True
wait["Qr"][msg.to] = True
wait["Qron"] = True
wait["Ghost"] = True
cl.sendText(msg.to,"All Protect Sudah Aktif Semua")
print wait["AutoCancel"]
print wait["AutoKick"]
print wait["Qr"]
else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["Allprotect off"]:
if msg.from_ in admin:
wait["AutoCancel"][msg.to] = False
wait["AutoCancelon"] = False
wait["inviteprotect"] = False
wait["joinkick"] = False
wait["AutoKick"][msg.to] = False
wait["AutoKickon"] = False
wait["Qr"][msg.to] = False
wait["Qron"] = False
wait["Ghost"] = False
cl.sendText(msg.to,"All Protect Sudah Di Nonaktifkan Semua")
print wait["AutoCancel"]
print wait["AutoKick"]
print wait["Qr"]
else:
#else:
cl.sendText(msg.to,"Only Admin")
elif msg.text in ["K on","Contact on"]:
wait["Contact"] = True
cl.sendText(msg.to,"Contact Sudah Aktif")
elif msg.text in ["K off","Contact off"]:
wait["Contact"] = False
cl.sendText(msg.to,"Contact Sudah Di Nonaktifkan")
elif msg.text in ["Alwaysread on"]:
wait["alwaysRead"] = True
cl.sendText(msg.to,"Always Read Sudah Aktif")
elif msg.text in ["Alwaysread off"]:
wait["alwaysRead"] = False
cl.sendText(msg.to,"Always Read Sudah Di Nonaktifkan")
elif msg.text in ["Sambutan on"]:
if wait["Sambutan"] == True:
if wait["lang"] == "JP":
cl.sendText(msg.to,"Sambutan Di Aktifkanヾ(*´∀`*)ノ")
else:
wait["Sambutan"] = True
wait["joinkick"] = False
if wait["lang"] == "JP":
cl.sendText(msg.to,"Sudah Onヽ(´▽`)/")
elif msg.text in ["Sambutan off"]:
if wait["Sambutan"] == False:
if wait["lang"] == "JP":
cl.sendText(msg.to,"Sambutan Di Nonaktifkan( ^∇^)")
else:
wait["Sambutan"] = False
if wait["lang"] == "JP":
cl.sendText(msg.to,"Sudah Off(p′︵‵。)")
elif "Sider on" in msg.text:
try:
del cctv['point'][msg.to]
del cctv['sidermem'][msg.to]
del cctv['cyduk'][msg.to]
except:
pass
cctv['point'][msg.to] = msg.id
cctv['sidermem'][msg.to] = ""
cctv['cyduk'][msg.to]=True
wait["Sider"] = True
cl.sendText(msg.to,"Siap On Cek Sider")
elif "Sider off" in msg.text:
if msg.to in cctv['point']:
cctv['cyduk'][msg.to]=False
wait["Sider"] = False
cl.sendText(msg.to, "Cek Sider Off")
else:
cl.sendText(msg.to, "Heh Belom Di Set")
elif msg.text in ["Status"]:
md = ""
if wait["Sambutan"] == True: md+="╠➩✔️ Sambutan : On\n"
else:md+="╠➩❌ Sambutan : Off\n"
if wait["joinkick"] == True: md+="╠➩✔️ Join Kick : On\n"
else:md+="╠➩❌ Join Kick : Off\n"
if wait["AutoJoin"] == True: md+="╠➩✔️ Auto Join : On\n"
else: md +="╠➩❌ Auto Join : Off\n"
if wait["AutoJoinCancel"] == True: md+="╠➩✔️ Auto Join Cancel : On\n"
else: md +="╠➩❌ Auto Join Cancel : Off\n"
if wait["Leave"] == True: md+="╠➩✔️ Leave : On\n"
else: md +="╠➩❌ Leave : Off\n"
if wait["Contact"] == True: md+="╠➩✔️ Info Contact : On\n"
else: md+="╠➩❌ Info Contact : Off\n"
if wait["AutoCancelon"] == True:md+="╠➩✔️ Auto Cancel : On\n"
else: md+= "╠➩❌ Auto Cancel : Off\n"
if wait["inviteprotect"] == True:md+="╠➩✔️ Invite Protect : On\n"
else: md+= "╠➩❌ Invite Protect : Off\n"
if wait["Qron"] == True: md+="╠➩✔️ Qr Protect : On\n"
else:md+="╠➩❌ Qr Protect : Off\n"
if wait["AutoKickon"] == True: md+="╠➩✔️ Auto Kick : On\n"
else:md+="╠➩❌ Auto Kick : Off\n"
if wait["Ghost"] == True: md+="╠➩✔️ Ghost : On\n"
else:md+="╠➩❌ Ghost : Off\n"
if wait["alwaysRead"] == True: md+="╠➩✔️ Always Read : On\n"
else:md+="╠➩❌ Always Read: Off\n"
if wait["detectMention"] == True: md+="╠➩✔️ Auto Respon : On\n"
else:md+="╠➩❌ Auto Respon : Off\n"
if wait["kickMention"] == True: md+="╠➩✔️ Auto Respon Kick : On\n"
else:md+="╠➩❌ Auto Respon Kick : Off\n"
if wait["Sider"] == True: md+="╠➩✔️ Auto Sider : On\n"
else:md+="╠➩❌ Auto Sider: Off\n"
if wait["Simi"] == True: md+="╠➩✔️ Simisimi : On\n"
else:md+="╠➩❌ Simisimi: Off\n"
cl.sendText(msg.to,"╔═════════════════════════\n""║ ☆☞ S T A T U S ☜☆\n""╠═════════════════════════\n"+md+"╚═════════════════════════")
elif msg.text in ["Gift","gift"]:
msg.contentType = 9
msg.contentMetadata={'PRDID': 'a0768339-c2d3-4189-9653-2909e9bb6f58',
'PRDTYPE': 'THEME',
'MSGTPL': '5'}
msg.text = None
cl.sendMessage(msg)
elif msg.text in ["All gift"]:
msg.contentType = 9
msg.contentMetadata={'PRDID': 'a0768339-c2d3-4189-9653-2909e9bb6f58',
'PRDTYPE': 'THEME',
'MSGTPL': '5'}
msg.text = None
ki.sendMessage(msg)
kk.sendMessage(msg)
kc.sendMessage(msg)
elif msg.text in ["TC1 Gift","TC1 gift"]:
msg.contentType = 9
msg.contentMetadata={'PRDID': '696d7046-843b-4ed0-8aac-3113ed6c0733',
'PRDTYPE': 'THEME',
'MSGTPL': '6'}
msg.text = None
ki.sendMessage(msg)
elif msg.text in ["TC2 Gift","TC2 gift"]:
msg.contentType = 9
msg.contentMetadata={'PRDID': '8fe8cdab-96f3-4f84-95f1-6d731f0e273e',
'PRDTYPE': 'THEME',
'MSGTPL': '7'}
msg.text = None
kk.sendMessage(msg)
elif msg.text in ["TC3 Gift","TC3 gift"]:
msg.contentType = 9
msg.contentMetadata={'PRDID': 'ae3d9165-fab2-4e70-859b-c14a9d4137c4',
'PRDTYPE': 'THEME',
'MSGTPL': '8'}
msg.text = None
kc.sendMessage(msg)
elif "Gift1 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift1 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '1',
'STKPKGID': '1380280'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift2 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift2 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '2',
'STKPKGID': '1360738'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift3 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift3 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '3',
'STKPKGID': '1395389'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift4 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift4 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '4',
'STKPKGID': '1329191'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift5 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift5 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '1',
'STKPKGID': '9057'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift6 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift6 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '2',
'STKPKGID': '9167'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift7 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift7 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '3',
'STKPKGID': '7334'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift8 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift8 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '1',
'STKPKGID': '1380280'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift9 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift9 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '4',
'STKPKGID': '1405277'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif "Gift10 " in msg.text:
msg.contentType = 13
nk0 = msg.text.replace("Gift10 ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
gs = cl.getGroup(msg.to)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
cl.sendText(msg.to,_name + " Check Your Gift")
msg.contentType = 9
msg.contentMetadata= {'PRDTYPE': 'STICKER',
'STKVER': '1',
'MSGTPL': '1',
'STKPKGID': '1296261'}
msg.to = target
msg.text = None
cl.sendMessage(msg)
except:
msg.contentMetadata = {'mid': target}
elif msg.text.lower() in ["wkwkwk","wkwk","hahaha","haha"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '100',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["hehehe","hehe"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '10',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["galau"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '9',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["you","kau","kamu"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '7',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["marah","hadeuh","hadeh"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '6',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["please","pliss","mohon","tolong"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '4',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["haa","haaa","kaget"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '3',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["lucu","ngakak","lol"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '110',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["hmm","hmmm"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '101',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["tidur"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '1',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["gemes"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '2',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["cantik","imut"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '5',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["nyanyi","lalala"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '11',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["gugup"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '8',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["ok","oke","okay","oce","okee","sip","siph"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '13',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["mantab","mantap","nice","keren"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '14',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["ngejek"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '15',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["nangis","sedih"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '16',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["woi","kampret"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '102',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text.lower() in ["huft"]:
msg.contentType = 7
msg.contentMetadata={'STKID': '104',
'STKPKGID': '1',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif msg.text in ["Tagall","Tag all"]:
group = cl.getGroup(msg.to)
nama = [contact.mid for contact in group.members]
cb = ""
cb2 = ""
strt = int(0)
akh = int(0)
for md in nama:
akh = akh + int(6)
cb += """{"S":"""+json.dumps(str(strt))+""","E":"""+json.dumps(str(akh))+""","M":"""+json.dumps(md)+"},"""
strt = strt + int(7)
akh = akh + 1
cb2 += "@nrik \n"
cb = (cb[:int(len(cb)-1)])
msg.contentType = 0
msg.text = cb2
msg.contentMetadata ={'MENTION':'{"MENTIONEES":['+cb+']}','EMTVER':'4'}
try:
cl.sendMessage(msg)
except Exception as error:
print error
elif msg.text in ["Setview","Setpoint","Cctv"]:
subprocess.Popen("echo '' > dataSeen/"+msg.to+".txt", shell=True, stdout=subprocess.PIPE)
cl.sendText(msg.to, "☆Checkpoint Checked☆")
print "Setview"
elif msg.text in ["Viewseen","Check","Ciduk","Cyduk"]:
lurkGroup = ""
dataResult, timeSeen, contacts, userList, timelist, recheckData = [], [], [], [], [], []
with open('dataSeen/'+msg.to+'.txt','r') as rr:
contactArr = rr.readlines()
for v in xrange(len(contactArr) -1,0,-1):
num = re.sub(r'\n', "", contactArr[v])
contacts.append(num)
pass
contacts = list(set(contacts))
for z in range(len(contacts)):
arg = contacts[z].split('|')
userList.append(arg[0])
timelist.append(arg[1])
uL = list(set(userList))
for ll in range(len(uL)):
try:
getIndexUser = userList.index(uL[ll])
timeSeen.append(time.strftime("%H:%M:%S", time.localtime(int(timelist[getIndexUser]) / 1000)))
recheckData.append(userList[getIndexUser])
except IndexError:
conName.append('nones')
pass
contactId = cl.getContacts(recheckData)
for v in range(len(recheckData)):
dataResult.append(contactId[v].displayName + ' ('+timeSeen[v]+')')
pass
if len(dataResult) > 0:
tukang = "╔═════════════════════════\n║ ☆☞ LIST VIEWERS ☜☆\n╠═════════════════════════\n╠➩"
grp = '\n╠➩ '.join(str(f) for f in dataResult)
total = '\n╠═════════════════════════\n╠➩ Total %i Viewers (%s)' % (len(dataResult), datetime.now().strftime('%H:%M:%S')) + "\n╚═════════════════════════"
cl.sendText(msg.to, "%s %s %s" % (tukang, grp, total))
subprocess.Popen("echo '' > dataSeen/"+msg.to+".txt", shell=True, stdout=subprocess.PIPE)
cl.sendText(msg.to, "☆Auto Checkpoint☆")
else:
cl.sendText(msg.to, "☆Belum Ada Viewers☆")
print "Viewseen"
elif "Kick " in msg.text:
if msg.from_ in admin:
if 'MENTION' in msg.contentMetadata.keys()!= None:
names = re.findall(r'@(\w+)', msg.text)
mention = ast.literal_eval(msg.contentMetadata['MENTION'])
mentionees = mention['MENTIONEES']
print mentionees
for mention in mentionees:
ki.kickoutFromGroup(msg.to,[mention['M']])
elif "Set member: " in msg.text:
if msg.from_ in admin:
jml = msg.text.replace("Set member: ","")
wait["memberscancel"] = int(jml)
cl.sendText(msg.to, "Jumlah minimal member telah di set : "+jml)
elif "Add all" in msg.text:
thisgroup = cl.getGroups([msg.to])
Mids = [contact.mid for contact in thisgroup[0].members]
mi_d = Mids[:33]
cl.findAndAddContactsByMids(mi_d)
cl.sendText(msg.to,"Success Add all")
elif msg.text in ["Invite"]:
wait["invite"] = True
cl.sendText(msg.to,"Send Contact")
elif msg.text in ["Auto like"]:
wait["likeOn"] = True
cl.sendText(msg.to,"Shere Post Kamu Yang Mau Di Like!")
elif msg.text in ["Steal contact"]:
wait["steal"] = True
cl.sendText(msg.to,"Send Contact")
elif msg.text in ["Giftbycontact"]:
wait["gift"] = True
cl.sendText(msg.to,"Send Contact")
elif "Recover" in msg.text:
thisgroup = cl.getGroups([msg.to])
Mids = [contact.mid for contact in thisgroup[0].members]
mi_d = Mids[:33]
cl.createGroup("Recover", mi_d)
cl.sendText(msg.to,"Success recover")
elif ("Gn: " in msg.text):
if msg.toType == 2:
X = cl.getGroup(msg.to)
X.name = msg.text.replace("Gn: ","")
cl.updateGroup(X)
else:
cl.sendText(msg.to,"It can't be used besides the group.")
elif "Kick: " in msg.text:
midd = msg.text.replace("Kick: ","")
kicker = [ki,kk,kc]
if midd not in admin:
random.choice(kicker).kickoutFromGroup(msg.to,[midd])
else:
cl.sendText(msg.to,"Admin Detected")
elif "Invite: " in msg.text:
midd = msg.text.replace("Invite: ","")
cl.findAndAddContactsByMid(midd)
ki.findAndAddContactsByMid(midd)
kk.findAndAddContactsByMid(midd)
kc.findAndAddContactsByMid(midd)
kr.findAndAddContactsByMid(midd)
random.choice(KAC).inviteIntoGroup(msg.to,[midd])
elif "Invite creator" in msg.text:
midd = "u71b6799e1c37868a871d442e67633182"
random.choice(KAC).inviteIntoGroup(msg.to,[midd])
elif msg.text in ["Welcome","welcome","Welkam","welkam","Wc","wc"]:
gs = cl.getGroup(msg.to)
cl.sendText(msg.to,"Selamat Datang Di "+ gs.name)
msg.contentType = 7
msg.contentMetadata={'STKID': '247',
'STKPKGID': '3',
'STKVER': '100'}
msg.text = None
cl.sendMessage(msg)
elif "Bc: " in msg.text:
bc = msg.text.replace("Bc: ","")
gid = cl.getGroupIdsJoined()
if msg.from_ in Creator:
for i in gid:
cl.sendText(i,"=======[BROADCAST]=======\n\n"+bc+"\n\nContact Me : line.me/ti/p/~a_ulul15")
cl.sendText(msg.to,"Success BC BosQ")
else:
cl.sendText(msg.to,"Khusus Admin")
elif msg.text in ["Cancel"]:
gid = cl.getGroupIdsInvited()
for i in gid:
cl.rejectGroupInvitation(i)
cl.sendText(msg.to,"All invitations have been refused")
elif msg.text in ["TC1 Cancel"]:
gid = ki.getGroupIdsInvited()
for i in gid:
ki.rejectGroupInvitation(i)
ki.sendText(msg.to,"All invitations have been refused")
elif msg.text in ["TC2 Cancel"]:
gid = kk.getGroupIdsInvited()
for i in gid:
kk.rejectGroupInvitation(i)
kk.sendText(msg.to,"All invitations have been refused")
elif msg.text in ["TC3 Cancel"]:
gid = kc.getGroupIdsInvited()
for i in gid:
kc.rejectGroupInvitation(i)
kc.sendText(msg.to,"All invitations have been refused")
elif msg.text in ["Gurl"]:
if msg.toType == 2:
x = cl.getGroup(msg.to)
if x.preventJoinByTicket == True:
x.preventJoinByTicket = False
cl.updateGroup(x)
gurl = cl.reissueGroupTicket(msg.to)
cl.sendText(msg.to,"line://ti/g/" + gurl)
else:
if wait["lang"] == "JP":
cl.sendText(msg.to,"Can't be used outside the group")
else:
cl.sendText(msg.to,"Not for use less than group")
elif msg.text in ["All join","Join all"]:
if msg.from_ in admin:
G = cl.getGroup(msg.to)
ginfo = cl.getGroup(msg.to)
G.preventJoinByTicket = False
cl.updateGroup(G)
invsend = 0
Ticket = cl.reissueGroupTicket(msg.to)
ki.acceptGroupInvitationByTicket(msg.to,Ticket)
time.sleep(0.2)
kk.acceptGroupInvitationByTicket(msg.to,Ticket)
time.sleep(0.2)
kc.acceptGroupInvitationByTicket(msg.to,Ticket)
time.sleep(0.2)
kr.acceptGroupInvitationByTicket(msg.to,Ticket)
time.sleep(0.2)
G = cl.getGroup(msg.to)
G.preventJoinByTicket = True
ki.updateGroup(G)
G.preventJoinByTicket(G)
ki.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["TC1 join"]:
if msg.from_ in admin:
X = cl.getGroup(msg.to)
X.preventJoinByTicket = False
cl.updateGroup(X)
invsend = 0
Ti = cl.reissueGroupTicket(msg.to)
ki.acceptGroupInvitationByTicket(msg.to,Ti)
G = kk.getGroup(msg.to)
G.preventJoinByTicket = True
ki.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["TC2 join"]:
if msg.from_ in admin:
X = cl.getGroup(msg.to)
X.preventJoinByTicket = False
cl.updateGroup(X)
invsend = 0
Ti = cl.reissueGroupTicket(msg.to)
kk.acceptGroupInvitationByTicket(msg.to,Ti)
G = ki.getGroup(msg.to)
G.preventJoinByTicket = True
kk.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["TC3 join"]:
if msg.from_ in admin:
G = cl.getGroup(msg.to)
ginfo = cl.getGroup(msg.to)
G.preventJoinByTicket = False
cl.updateGroup(G)
invsend = 0
Ticket = cl.reissueGroupTicket(msg.to)
kc.acceptGroupInvitationByTicket(msg.to,Ticket)
G.preventJoinByTicket = True
kc.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["TC4 join"]:
if msg.from_ in admin:
G = cl.getGroup(msg.to)
ginfo = cl.getGroup(msg.to)
G.preventJoinByTicket = False
cl.updateGroup(G)
invsend = 0
Ticket = cl.reissueGroupTicket(msg.to)
kr.acceptGroupInvitationByTicket(msg.to,Ticket)
G.preventJoinByTicket = True
kr.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["Ghost join"]:
if msg.from_ in admin:
G = cl.getGroup(msg.to)
ginfo = cl.getGroup(msg.to)
G.preventJoinByTicket = False
cl.updateGroup(G)
invsend = 0
Ticket = cl.reissueGroupTicket(msg.to)
km.acceptGroupInvitationByTicket(msg.to,Ticket)
G.preventJoinByTicket = True
km.updateGroup(G)
else:
cl.sendText(msg.to,"Sape lu!")
elif msg.text in ["timeline"]:
try:
url = cl.activity(limit=5)
cl.sendText(msg.to,url['result']['posts'][0]['postInfo']['postId'])
except Exception as E:
print E
elif msg.text in ["Bye all"]:
if wait["Leave"] == True:
ki.leaveGroup(msg.to)
kk.leaveGroup(msg.to)
kc.leaveGroup(msg.to)
kr.leaveGroup(msg.to)
else:
cl.sendText(msg.to,"Leavenya Belum On")
elif msg.text in ["@bye","@Bye"]:
if wait["Leave"] == True:
cl.leaveGroup(msg.to)
wait["Leave"] = False
else:
cl.sendText(msg.to,"Bilang Dulu Sama Admin Ku")
elif msg.text in ["Absen"]:
cl.sendText(msg.to,"Pasukan Absen!!")
ki.sendText(msg.to,"TC1 Hadiir \(ˆ▿ˆ)/")
kk.sendText(msg.to,"TC2 Hadiir \(ˆ▿ˆ)/")
kc.sendText(msg.to,"TC3 Hadiir \(ˆ▿ˆ)/")
kr.sendText(msg.to,"Hadiir Semua Kapten \(ˆ▿ˆ)/")
elif msg.text.lower() in ["respon"]:
cl.sendText(msg.to,responsename)
ki.sendText(msg.to,responsename2)
kk.sendText(msg.to,responsename3)
kc.sendText(msg.to,responsename4)
kr.sendText(msg.to,responsename5)
elif msg.text in ["Sp","Speed","speed"]:
start = time.time()
print("Speed")
elapsed_time = time.time() - start
cl.sendText(msg.to, "Tunggu Bentaar BOS....")
cl.sendText(msg.to, "%sseconds" % (elapsed_time))
elif msg.text in ["Speed test"]:
start = time.time()
cl.sendText(msg.to, "Tunggu Bentaar BOS......")
elapsed_time = time.time() - start
cl.sendText(msg.to, "%sseconds" % (elapsed_time))
elif "Nk: " in msg.text:
if msg.from_ in Creator:
X = cl.getGroup(msg.to)
X.preventJoinByTicket = False
cl.updateGroup(X)
invsend = 0
Ti = cl.reissueGroupTicket(msg.to)
kr.acceptGroupInvitationByTicket(msg.to,Ti)
G = kk.getGroup(msg.to)
G.preventJoinByTicket = True
kk.updateGroup(G)
nk0 = msg.text.replace("Nk: ","")
nk1 = nk0.lstrip()
nk2 = nk1.replace("@","")
nk3 = nk2.rstrip()
_name = nk3
targets = []
for s in X.members:
if _name in s.displayName:
targets.append(s.mid)
if targets == []:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
if target not in admin:
kr.kickoutFromGroup(msg.to,[target])
kr.leaveGroup(msg.to)
ki.sendText(msg.to,"Succes BosQ")
kk.sendText(msg.to,"Pakyu~")
else:
cl.sendText(msg.to,"Admin Detected")
else:
cl.sendText(msg.to,"Lu sape!")
elif msg.text in ["Ban"]:
if msg.from_ in admin:
wait["wblacklist"] = True
ki.sendText(msg.to,"send contact")
elif msg.text in ["Unban"]:
if msg.from_ in admin:
wait["dblacklist"] = True
ki.sendText(msg.to,"send contact")
elif "Ban @" in msg.text:
if msg.from_ in admin:
if msg.toType == 2:
print "@Ban by mention"
_name = msg.text.replace("Ban @","")
_nametarget = _name.rstrip(' ')
gs = ki.getGroup(msg.to)
gs = kk.getGroup(msg.to)
gs = kc.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
kc.sendText(msg.to,"Not found")
else:
for target in targets:
if target not in admin:
try:
wait["blacklist"][target] = True
f=codecs.open('st2__b.json','w','utf-8')
json.dump(wait["blacklist"], f, sort_keys=True, indent=4,ensure_ascii=False)
ki.sendText(msg.to,"Succes BosQ")
except:
ki.sendText(msg.to,"Error")
else:
cl.sendText(msg.to,"Admin Detected~")
elif msg.text in ["Banlist","Ban list"]:
if msg.from_ in admin:
if wait["blacklist"] == {}:
random.choice(KAC).sendText(msg.to,"Tidak Ada")
else:
mc = ""
for mi_d in wait["blacklist"]:
mc += "->" +cl.getContact(mi_d).displayName + "\n"
random.choice(KAC).sendText(msg.to,"===[Blacklist User]===\n"+mc)
elif "Unban @" in msg.text:
if msg.toType == 2:
print "@Unban by mention"
if msg.from_ in admin:
_name = msg.text.replace("Unban @","")
_nametarget = _name.rstrip(' ')
gs = ki.getGroup(msg.to)
gs = kk.getGroup(msg.to)
gs = kc.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
kk.sendText(msg.to,"Not found")
else:
for target in targets:
try:
del wait["blacklist"][target]
f=codecs.open('st2__b.json','w','utf-8')
json.dump(wait["blacklist"], f, sort_keys=True, indent=4,ensure_ascii=False)
ki.sendText(msg.to,"Succes BosQ")
except:
ki.sendText(msg.to,"Succes BosQ")
elif msg.text.lower() == 'clear ban':
if msg.from_ in admin:
wait["blacklist"] = {}
cl.sendText(msg.to,"ヽ( ^ω^)ノ└ ❉Unbanned All Success❉ ┐")
elif msg.text.lower() in ["sayang","chucky"]:
ki.sendText(msg.to,"Apa Sayang :*")
elif msg.text in ["Kill ban"]:
if msg.from_ in admin:
if msg.toType == 2:
group = cl.getGroup(msg.to)
gMembMids = [contact.mid for contact in group.members]
matched_list = []
for tag in wait["blacklist"]:
matched_list+=filter(lambda str: str == tag, gMembMids)
if matched_list == []:
ki.sendText(msg.to,"There was no blacklist user")
return
for jj in matched_list:
random.choice(KAC).kickoutFromGroup(msg.to,[jj])
ki.sendText(msg.to,"Blacklist emang pantas tuk di usir")
else:
cl.sendText(msg.to, "Khusus creator")
elif msg.text in ["Kill"]:
if msg.toType == 2:
if msg.from_ in admin:
group = ki.getGroup(msg.to)
gMembMids = [contact.mid for contact in group.members]
matched_list = []
for tag in wait["blacklist"]:
matched_list+=filter(lambda str: str == tag, gMembMids)
if matched_list == []:
kk.sendText(msg.to,"Fuck You")
kc.sendText(msg.to,"Fuck You")
return
for jj in matched_list:
try:
klist=[ki,kk,kc]
kicker=random.choice(klist)
kicker.kickoutFromGroup(msg.to,[jj])
print (msg.to,[jj])
except:
pass
elif "Kickall" == msg.text:
if msg.from_ in Creator:
if msg.toType == 2:
print "Kick all member"
_name = msg.text.replace("Kickall","")
gs = ki.getGroup(msg.to)
gs = kk.getGroup(msg.to)
gs = kc.getGroup(msg.to)
ki.sendText(msg.to,"Sampai jumpaa~")
kc.sendText(msg.to,"Dadaaah~")
targets = []
for g in gs.members:
if _name in g.displayName:
targets.append(g.mid)
if targets == []:
ki.sendText(msg.to,"Not found.")
else:
for target in targets:
if target not in admin:
try:
klist=[ki,kk,kc]
kicker=random.choice(klist)
kicker.kickoutFromGroup(msg.to,[target])
print (msg.to,[g.mid])
except Exception as e:
cl.sendText(msg.to,str(e))
cl.inviteIntoGroup(msg.to, targets)
elif msg.text in ["Bot restart","Reboot"]:
if msg.from_ in Creator:
cl.sendText(msg.to, "Bot Has Been Restarted...")
restart_program()
print "@Restart"
else:
cl.sendText(msg.to, "No Access")
elif msg.text in ["Turn off"]:
if msg.from_ in Creator:
try:
import sys
sys.exit()
except:
pass
elif 'Crash' in msg.text:
if msg.from_ in Creator:
msg.contentType = 13
msg.contentMetadata = {'mid': "NADYA,'"}
cl.sendMessage(msg)
elif "Kapten copy @" in msg.text:
print "[COPY] Ok"
_name = msg.text.replace("Kapten copy @","")
_nametarget = _name.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to, "Not Found...")
else:
for target in targets:
try:
cl.CloneContactProfile(target)
cl.sendText(msg.to, "Copied (^_^)")
except Exception as e:
print e
elif "TC1 copy @" in msg.text:
print "[COPY] Ok"
_name = msg.text.replace("TC1 copy @","")
_nametarget = _name.rstrip(' ')
gs = ki.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
ki.sendText(msg.to, "Not Found...")
else:
for target in targets:
try:
ki.CloneContactProfile(target)
ki.sendText(msg.to, "Copied (^_^)")
except Exception as e:
print e
elif "TC2 copy @" in msg.text:
print "[COPY] Ok"
_name = msg.text.replace("TC2 copy @","")
_nametarget = _name.rstrip(' ')
gs = kk.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
kk.sendText(msg.to, "Not Found...")
else:
for target in targets:
try:
kk.CloneContactProfile(target)
kk.sendText(msg.to, "Copied (^_^)")
except Exception as e:
print e
elif "TC3 copy @" in msg.text:
print "[COPY] Ok"
_name = msg.text.replace("TC3 copy @","")
_nametarget = _name.rstrip(' ')
gs = kc.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
kc.sendText(msg.to, "Not Found...")
else:
for target in targets:
try:
kc.CloneContactProfile(target)
kc.sendText(msg.to, "Copied (^_^)")
except Exception as e:
print e
elif "TC4 copy @" in msg.text:
print "[COPY] Ok"
_name = msg.text.replace("TC4 copy @","")
_nametarget = _name.rstrip(' ')
gs = kr.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
kr.sendText(msg.to, "Not Found...")
else:
for target in targets:
try:
kr.CloneContactProfile(target)
kr.sendText(msg.to, "Copied (^_^)")
except Exception as e:
print e
elif msg.text in ["Backup all"]:
try:
ki.updateDisplayPicture(backup2.pictureStatus)
ki.updateProfile(backup2)
kk.updateDisplayPicture(backup3.pictureStatus)
kk.updateProfile(backup3)
kc.updateDisplayPicture(backup4.pictureStatus)
kc.updateProfile(backup4)
kr.updateDisplayPicture(backup5.pictureStatus)
kr.updateProfile(backup5)
cl.updateDisplayPicture(backup1.pictureStatus)
cl.updateProfile(backup1)
cl.sendText(msg.to, "All Done (^_^)")
except Exception as e:
cl.sendText(msg.to, str(e))
elif "/musik " in msg.text:
songname = msg.text.replace("/musik ","")
params = {"songname": songname}
r = requests.get('http://ide.fdlrcn.com/workspace/yumi-apis/joox?' + urllib.urlencode(params))
data = r.text
data = json.loads(data)
for song in data:
abc = song[3].replace('https://','http://')
cl.sendText(msg.to, "Title : " + song[0] + "\nLength : " + song[1] + "\nLink download : " + song[4])
cl.sendText(msg.to, "Lagu " + song[0] + "\nSedang Di Prosses... Tunggu Sebentar ^_^ ")
cl.sendAudioWithURL(msg.to,abc)
cl.sendText(msg.to, "Selamat Mendengarkan Lagu " + song[0])
elif '/lirik ' in msg.text.lower():
try:
songname = msg.text.lower().replace('/lirik ','')
params = {'songname': songname}
r = requests.get('http://ide.fdlrcn.com/workspace/yumi-apis/joox?' + urllib.urlencode(params))
data = r.text
data = json.loads(data)
for song in data:
hasil = 'Lyric Lagu ('
hasil += song[0]
hasil += ')\n\n'
hasil += song[5]
cl.sendText(msg.to, hasil)
except Exception as wak:
cl.sendText(msg.to, str(wak))
elif "/musrik " in msg.text:
songname = msg.text.replace("/musrik ","")
params = {"songname": songname}
r = requests.get('http://ide.fdlrcn.com/workspace/yumi-apis/joox?' + urllib.urlencode(params))
data = r.text
data = json.loads(data)
for song in data:
abc = song[3].replace('https://','http://')
hasil = 'Lyric Lagu ('
hasil += song[0]
hasil += ')\n\n'
hasil += song[5]
cl.sendText(msg.to, "Lagu " + song[0] + "\nSedang Di Prosses... Tunggu Sebentar ^_^ ")
cl.sendAudioWithURL(msg.to,abc)
cl.sendText(msg.to, "Title : " + song[0] + "\nLength : " + song[1] + "\nLink download : " + song[4] +"\n\n" + hasil)
cl.sendText(msg.to, "Selamat Mendengarkan Lagu " + song[0])
elif "Fancytext: " in msg.text:
txt = msg.text.replace("Fancytext: ", "")
cl.kedapkedip(msg.to,txt)
print "[Command] Kedapkedip"
elif "cover @" in msg.text:
if msg.toType == 2:
cover = msg.text.replace("cover @","")
_nametarget = cover.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to,"Not found")
else:
for target in targets:
try:
h = cl.channel.getHome(target)
objId = h["result"]["homeInfo"]["objectId"]
cl.sendImageWithURL(msg.to,"http://dl.profile.line-cdn.net/myhome/c/download.nhn?userid=" + target + "&oid=" + objId)
except Exception as error:
print error
cl.sendText(msg.to,"Upload image failed.")
elif "Cover @" in msg.text:
if msg.toType == 2:
cover = msg.text.replace("Cover @","")
_nametarget = cover.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to,"Not found")
else:
for target in targets:
try:
h = cl.channel.getHome(target)
objId = h["result"]["homeInfo"]["objectId"]
cl.sendImageWithURL(msg.to,"http://dl.profile.line-cdn.net/myhome/c/download.nhn?userid=" + target + "&oid=" + objId)
except Exception as error:
print error
cl.sendText(msg.to,"Upload image failed.")
elif "pp @" in msg.text:
if msg.toType == 2:
cover = msg.text.replace("pp @","")
_nametarget = cover.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to,"Not found")
else:
for target in targets:
try:
h = cl.getContact(target)
cl.sendImageWithURL(msg.to,"http://dl.profile.line-cdn.net/" + h.pictureStatus)
except Exception as error:
print error
cl.sendText(msg.to,"Upload image failed.")
elif "Pp @" in msg.text:
if msg.toType == 2:
cover = msg.text.replace("Pp @","")
_nametarget = cover.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to,"Not found")
else:
for target in targets:
try:
h = cl.getContact(target)
cl.sendImageWithURL(msg.to,"http://dl.profile.line-cdn.net/" + h.pictureStatus)
except Exception as error:
print error
cl.sendText(msg.to,"Upload image failed.")
elif msg.text.lower() in ["van","yog","wan","gong","tep","pap creator"]:
link = ["http://dl.profile.line-cdn.net/0hbPvoismJPX9LFhHY8ztCKHdTMxI8OCw3JXclGzwRYBpgci99fyV2GzwUY01icXl5J3EnHjxBakxj"]
pilih = random.choice(link)
ki.sendImageWithURL(msg.to,pilih)
elif msg.text.lower() in ["van","yog","wan","gong","tep","pap owner","pap creator"]:
link = ["http://dl.profile.line-cdn.net/0hbPvoismJPX9LFhHY8ztCKHdTMxI8OCw3JXclGzwRYBpgci99fyV2GzwUY01icXl5J3EnHjxBakxj"]
pilih = random.choice(link)
ki.sendImageWithURL(msg.to,pilih)
elif "Spam: " in msg.text:
bctxt = msg.text.replace("Spam: ", "")
t = 10
while(t):
random.choice(KAC).sendText(msg.to, (bctxt))
t-=1
elif "Scbc " in msg.text:
bctxt = msg.text.replace("Scbc ", "")
orang = cl.getAllContactIds()
t = 20
for manusia in orang:
while(t):
cl.sendText(manusia, (bctxt))
t-=1
elif "Cbc " in msg.text:
broadcasttxt = msg.text.replace("Cbc ", "")
orang = cl.getAllContactIds()
for manusia in orang:
cl.sendText(manusia, (broadcasttxt))
elif '/ig ' in msg.text.lower():
try:
instagram = msg.text.lower().replace("/ig ","")
html = requests.get('https://www.instagram.com/' + instagram + '/?')
soup = BeautifulSoup(html.text, 'html.parser')
data = soup.find_all('meta', attrs={'property':'og:description'})
text = data[0].get('content').split()
data1 = soup.find_all('meta', attrs={'property':'og:image'})
text1 = data1[0].get('content').split()
nadya = text1[0].replace("s150x150/","")
user = "Name: " + text[-2] + "\n"
user1 = "Username: " + text[-1] + "\n"
followers = "Followers: " + text[0] + "\n"
following = "Following: " + text[2] + "\n"
post = "Post: " + text[4] + "\n"
link = "Link: " + "https://www.instagram.com/" + instagram
detail = "========INSTAGRAM INFO ========\n"
details = "\n========INSTAGRAM INFO ========"
cl.sendText(msg.to, detail + user + user1 + followers + following + post + link + details)
cl.sendImageWithURL(msg.to, nadya)
except Exception as njer:
cl.sendText(msg.to, str(njer))
elif "Checkig " in msg.text:
separate = msg.text.split(" ")
user = msg.text.replace(separate[0] + " ","")
if user.startswith("@"):
user = user.replace("@","")
profile = "https://www.instagram.com/" + user
with requests.session() as x:
x.headers['user-agent'] = 'Mozilla/5.0'
end_cursor = ''
for count in range(1, 999):
print('PAGE: ', count)
r = x.get(profile, params={'max_id': end_cursor})
data = re.search(r'window._sharedData = (\{.+?});</script>', r.text).group(1)
j = json.loads(data)
for node in j['entry_data']['ProfilePage'][0]['user']['media']['nodes']:
if node['is_video']:
page = 'https://www.instagram.com/p/' + node['code']
r = x.get(page)
url = re.search(r'"video_url": "([^"]+)"', r.text).group(1)
print(url)
cl.sendVideoWithURL(msg.to,url)
else:
print (node['display_src'])
cl.sendImageWithURL(msg.to,node['display_src'])
end_cursor = re.search(r'"end_cursor": "([^"]+)"', r.text).group(1)
elif 'Youtubelink: ' in msg.text:
try:
textToSearch = (msg.text).replace('Youtube ', "").strip()
query = urllib.quote(textToSearch)
url = "https://www.youtube.com/results?search_query=" + query
response = urllib2.urlopen(url)
html = response.read()
soup = BeautifulSoup(html, "html.parser")
results = soup.find(attrs={'class':'yt-uix-tile-link'})
cl.sendText(msg.to,'https://www.youtube.com' + results['href'])
except:
cl.sendText(msg.to,"Could not find it")
elif 'Youtubevideo: ' in msg.text:
try:
textToSearch = (msg.text).replace('Youtubevideo: ', "").strip()
query = urllib.quote(textToSearch)
url = "https://www.youtube.com/results?search_query=" + query
response = urllib2.urlopen(url)
html = response.read()
soup = BeautifulSoup(html, "html.parser")
results = soup.find(attrs={'class': 'yt-uix-tile-link'})
cl.sendVideoWithURL(msg.to,'https://www.youtube.com' + results['href'])
except:
cl.sendText(msg.to, "Could not find it")
elif "Say-id " in msg.text:
say = msg.text.replace("Say-id ","")
lang = 'id'
tts = gTTS(text=say, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "Say-en " in msg.text:
say = msg.text.replace("Say-en ","")
lang = 'en'
tts = gTTS(text=say, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "Say-jp " in msg.text:
say = msg.text.replace("Say-jp ","")
lang = 'ja'
tts = gTTS(text=say, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "Say welcome" in msg.text:
gs = cl.getGroup(msg.to)
say = msg.text.replace("Say welcome","Selamat Datang Di "+ gs.name)
lang = 'id'
tts = gTTS(text=say, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif msg.text.lower() in ["hi","hai","halo","hallo"]:
beb = "Hi Sayang 😘 " +cl.getContact(msg.from_).displayName + " starry heart"
kr.sendText(msg.to,beb)
elif "playstore " in msg.text.lower():
tob = msg.text.lower().replace("playstore ","")
cl.sendText(msg.to,"Sedang Mencari...")
cl.sendText(msg.to,"Title : "+tob+"\nSource : Google Play\nLink : https://play.google.com/store/search?q=" + tob)
cl.sendText(msg.to,"Tuh Linknya Kak (^_^)")
elif "Mid @" in msg.text:
_name = msg.text.replace("Mid @","")
_nametarget = _name.rstrip(' ')
gs = cl.getGroup(msg.to)
for g in gs.members:
if _nametarget == g.displayName:
random.choice(KAC).sendText(msg.to, g.mid)
else:
pass
elif "/bio " in msg.text:
string = msg.text.replace("/bio ","")
if len(string.decode('utf-8')) <= 500:
profile = cl.getProfile()
profile.statusMessage = string
cl.updateProfile(profile)
ki.updateProfile(profile)
kk.updateProfile(profile)
kc.updateProfile(profile)
kr.updateProfile(profile)
cl.sendText(msg.to,"All Done")
elif "/cnkapten" in msg.text:
if msg.from_ in Creator:
string = msg.text.replace("/cnkapten","Mi Kapten")
if len(string.decode('utf-8')) <= 5000:
profile = cl.getProfile()
profile.displayName = string
cl.updateProfile(profile)
cl.sendText(msg.to,"Done")
elif "/cntc1" in msg.text:
if msg.from_ in Creator:
string = msg.text.replace("/cntc1","Mi TC1")
if len(string.decode('utf-8')) <= 5000:
profile = ki.getProfile()
profile.displayName = string
ki.updateProfile(profile)
ki.sendText(msg.to,"Done")
elif "/cntc2" in msg.text:
if msg.from_ in Creator:
string = msg.text.replace("/cntc2","Mi TC2")
if len(string.decode('utf-8')) <= 5000:
profile = kk.getProfile()
profile.displayName = string
kk.updateProfile(profile)
kk.sendText(msg.to,"Done")
elif "/cntc3" in msg.text:
if msg.from_ in Creator:
string = msg.text.replace("/cntc3","Mi TC3")
if len(string.decode('utf-8')) <= 5000:
profile = kc.getProfile()
profile.displayName = string
kc.updateProfile(profile)
kc.sendText(msg.to,"Done")
elif "/cntc4" in msg.text:
if msg.from_ in Creator:
string = msg.text.replace("/cntc4","Mi TC4")
if len(string.decode('utf-8')) <= 5000:
profile = cl.getProfile()
profile.displayName = string
kr.updateProfile(profile)
kr.sendText(msg.to,"Done")
elif "Ulti " in msg.text:
if msg.from_ in Creator:
ulti0 = msg.text.replace("Ulti ","")
ulti1 = ulti0.rstrip()
ulti2 = ulti1.replace("@","")
ulti3 = ulti2.rstrip()
_name = ulti3
gs = cl.getGroup(msg.to)
ginfo = cl.getGroup(msg.to)
gs.preventJoinByTicket = False
cl.updateGroup(gs)
invsend = 0
Ticket = cl.reissueGroupTicket(msg.to)
km.acceptGroupInvitationByTicket(msg.to,Ticket)
time.sleep(0.2)
targets = []
for s in gs.members:
if _name in s.displayName:
targets.append(s.mid)
if targets ==[]:
sendMessage(msg.to,"user does not exist")
pass
else:
for target in targets:
try:
km.kickoutFromGroup(msg.to,[target])
km.leaveGroup(msg.to)
print (msg.to,[g.mid])
except:
km.sendText(msg.t,"Ter ELIMINASI....")
km.sendText(msg.to,"WOLES brooo....!!!")
km.leaveGroup(msg.to)
gs = cl.getGroup(msg.to)
gs.preventJoinByTicket = True
cl.updateGroup(gs)
gs.preventJoinByTicket(gs)
cl.updateGroup(gs)
elif msg.text.lower() in ["mymid","myid"]:
middd = "Name : " +cl.getContact(msg.from_).displayName + "\nMid : " +msg.from_
kr.sendText(msg.to,middd)
elif msg.text.lower() in ["me"]:
msg.contentType = 13
msg.contentMetadata = {'mid': msg.from_}
cl.sendMessage(msg)
elif "/apakah " in msg.text:
apk = msg.text.replace("/apakah ","")
rnd = ["Ya","Tidak","Bisa Jadi","Mungkin"]
p = random.choice(rnd)
lang = 'id'
tts = gTTS(text=p, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "/hari " in msg.text:
apk = msg.text.replace("/hari ","")
rnd = ["Senin","Selasa","Rabu","Kamis","Jumat","Sabtu","Minggu"]
p = random.choice(rnd)
lang = 'id'
tts = gTTS(text=p, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "/berapa " in msg.text:
apk = msg.text.replace("/berapa ","")
rnd = ['10%','20%','30%','40%','50%','60%','70%','80%','90%','100%','0%']
p = random.choice(rnd)
lang = 'id'
tts = gTTS(text=p, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "/berapakah " in msg.text:
apk = msg.text.replace("/berapakah ","")
rnd = ['1','2','3','4','5','6','7','8','9','10','Tidak Ada']
p = random.choice(rnd)
lang = 'id'
tts = gTTS(text=p, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif "/kapan " in msg.text:
apk = msg.text.replace("/kapan ","")
rnd = ["kapan kapan","besok","satu abad lagi","Hari ini","Tahun depan","Minggu depan","Bulan depan","Sebentar lagi","Tidak Akan Pernah"]
p = random.choice(rnd)
lang = 'id'
tts = gTTS(text=p, lang=lang)
tts.save("hasil.mp3")
cl.sendAudio(msg.to,"hasil.mp3")
elif msg.text in ["Simisimi on","Simisimi:on"]:
settings["simiSimi"][msg.to] = True
wait["Simi"] = True
cl.sendText(msg.to," Simisimi Di Aktifkan")
elif msg.text in ["Simisimi off","Simisimi:off"]:
settings["simiSimi"][msg.to] = False
wait["Simi"] = False
cl.sendText(msg.to,"Simisimi Di Nonaktifkan")
elif "Image " in msg.text:
search = msg.text.replace("Image ","")
url = 'https://www.google.com/search?espv=2&biw=1366&bih=667&tbm=isch&oq=kuc&aqs=mobile-gws-lite.0.0l5&q=' + search
raw_html = (download_page(url))
items = []
items = items + (_images_get_all_items(raw_html))
path = random.choice(items)
print path
try:
cl.sendImageWithURL(msg.to,path)
except:
pass
elif "Youtubesearch: " in msg.text:
query = msg.text.replace("Youtube ","")
with requests.session() as s:
s.headers['user-agent'] = 'Mozilla/5.0'
url = 'http://www.youtube.com/results'
params = {'search_query': query}
r = s.get(url, params=params)
soup = BeautifulSoup(r.content, 'html.parser')
hasil = ""
for a in soup.select('.yt-lockup-title > a[title]'):
if '&list=' not in a['href']:
hasil += ''.join((a['title'],'\nUrl : http://www.youtube.com' + a['href'],'\n\n'))
cl.sendText(msg.to,hasil)
print '[Command] Youtube Search'
elif "Tr-id " in msg.text:
isi = msg.text.replace("Tr-id ","")
translator = Translator()
hasil = translator.translate(isi, dest='id')
A = hasil.text
A = A.encode('utf-8')
cl.sendText(msg.to, A)
elif "Tr-en " in msg.text:
isi = msg.text.replace("Tr-en ","")
translator = Translator()
hasil = translator.translate(isi, dest='en')
A = hasil.text
A = A.encode('utf-8')
cl.sendText(msg.to, A)
elif "Tr-th " in msg.text:
isi = msg.text.replace("Tr-th ","")
translator = Translator()
hasil = translator.translate(isi, dest='th')
A = hasil.text
A = A.encode('utf-8')
cl.sendText(msg.to, A)
elif "Id@en" in msg.text:
bahasa_awal = 'id'
bahasa_tujuan = 'en'
kata = msg.text.replace("Id@en ","")
url = 'https://translate.google.com/m?sl=%s&tl=%s&ie=UTF-8&prev=_m&q=%s' % (bahasa_awal, bahasa_tujuan, kata.replace(" ", "+"))
agent = {'User-Agent':'Mozilla/5.0'}
cari_hasil = 'class="t0">'
request = urllib2.Request(url, headers=agent)
page = urllib2.urlopen(request).read()
result = page[page.find(cari_hasil)+len(cari_hasil):]
result = result.split("<")[0]
cl.sendText(msg.to,"----Dari Indonesia----\n" + "" + kata + "\n\n----Ke Inggris----\n" + "" + result)
elif "En@id" in msg.text:
bahasa_awal = 'en'
bahasa_tujuan = 'id'
kata = msg.text.replace("En@id ","")
url = 'https://translate.google.com/m?sl=%s&tl=%s&ie=UTF-8&prev=_m&q=%s' % (bahasa_awal, bahasa_tujuan, kata.replace(" ", "+"))
agent = {'User-Agent':'Mozilla/5.0'}
cari_hasil = 'class="t0">'
request = urllib2.Request(url, headers=agent)
page = urllib2.urlopen(request).read()
result = page[page.find(cari_hasil)+len(cari_hasil):]
result = result.split("<")[0]
cl.sendText(msg.to,"----Dari Inggris----\n" + "" + kata + "\n\n----Ke Indonesia----\n" + "" + result)
elif "Id@th" in msg.text:
bahasa_awal = 'id'
bahasa_tujuan = 'th'
kata = msg.text.replace("Id@en ","")
url = 'https://translate.google.com/m?sl=%s&tl=%s&ie=UTF-8&prev=_m&q=%s' % (bahasa_awal, bahasa_tujuan, kata.replace(" ", "+"))
agent = {'User-Agent':'Mozilla/5.0'}
cari_hasil = 'class="t0">'
request = urllib2.Request(url, headers=agent)
page = urllib2.urlopen(request).read()
result = page[page.find(cari_hasil)+len(cari_hasil):]
result = result.split("<")[0]
cl.sendText(msg.to,"----Dari Indonesia----\n" + "" + kata + "\n\n----Ke Thailand----\n" + "" + result)
elif "Th@id" in msg.text:
bahasa_awal = 'th'
bahasa_tujuan = 'id'
kata = msg.text.replace("Id@en ","")
url = 'https://translate.google.com/m?sl=%s&tl=%s&ie=UTF-8&prev=_m&q=%s' % (bahasa_awal, bahasa_tujuan, kata.replace(" ", "+"))
agent = {'User-Agent':'Mozilla/5.0'}
cari_hasil = 'class="t0">'
request = urllib2.Request(url, headers=agent)
page = urllib2.urlopen(request).read()
result = page[page.find(cari_hasil)+len(cari_hasil):]
result = result.split("<")[0]
cl.sendText(msg.to,"----Dari Thailand----\n" + "" + kata + "\n\n----Ke Indonesia----\n" + "" + result)
elif msg.text in ["Friendlist"]:
contactlist = cl.getAllContactIds()
kontak = cl.getContacts(contactlist)
num=1
msgs="═════════List Friend═════════"
for ids in kontak:
msgs+="\n[%i] %s" % (num, ids.displayName)
num=(num+1)
msgs+="\n═════════List Friend═════════\n\nTotal Friend : %i" % len(kontak)
cl.sendText(msg.to, msgs)
elif msg.text in ["Memlist"]:
kontak = cl.getGroup(msg.to)
group = kontak.members
num=1
msgs="═════════List Member═�����═══════-"
for ids in group:
msgs+="\n[%i] %s" % (num, ids.displayName)
num=(num+1)
msgs+="\n═════════List Member═════════\n\nTotal Members : %i" % len(group)
cl.sendText(msg.to, msgs)
elif msg.text in ["Spam"]:
if msg.from_ in admin:
cl.sendText(msg.to,"Aku belum mandi")
ki.sendText(msg.to,"Tak tun tuang")
kk.sendText(msg.to,"Tak tun tuang")
cl.sendText(msg.to,"Tapi masih cantik juga")
ki.sendText(msg.to,"Tak tun tuang")
kk.sendText(msg.to,"Tak tun tuang")
cl.sendText(msg.to,"apalagi kalau sudah mandi")
ki.sendText(msg.to,"Tak tun tuang")
kk.sendText(msg.to,"Pasti cantik sekali")
cl.sendText(msg.to,"yiha")
ki.sendText(msg.to,"Kalau orang lain melihatku")
kk.sendText(msg.to,"Tak tun tuang")
cl.sendText(msg.to,"Badak aku taba bana")
ki.sendText(msg.to,"Tak tun tuang")
kk.sendText(msg.to,"Tak tuntuang")
cl.sendText(msg.to,"Tapi kalau langsuang diidu")
ki.sendText(msg.to,"Tak tun tuang")
kk.sendText(msg.to,"Atagfirullah baunya")
cl.sendText(msg.to,"Males lanjutin ah")
ki.sendText(msg.to,"Sepi bat")
kk.sendText(msg.to,"Iya sepi udah udah")
cl.sendText(msg.to,"Gaada yang denger juga kita nyanyi")
ki.sendText(msg.to,"Nah")
kk.sendText(msg.to,"Mending gua makan dulu")
cl.sendText(msg.to,"Siyap")
ki.sendText(msg.to,"Okeh")
kk.sendText(msg.to,"Katanya owner kita Jomblo ya")
cl.sendText(msg.to,"Iya emang")
ki.sendText(msg.to,"Denger denger si lagi nyari pacar doi")
kk.sendText(msg.to,"Udah ah gosip mulu doain aja biar dapet")
elif "Getvid @" in msg.text:
print "[Command]dp executing"
_name = msg.text.replace("Getvid @","")
_nametarget = _name.rstrip(' ')
gs = cl.getGroup(msg.to)
targets = []
for g in gs.members:
if _nametarget == g.displayName:
targets.append(g.mid)
if targets == []:
cl.sendText(msg.to,"Contact not found")
else:
for target in targets:
try:
contact = cl.getContact(target)
path = "http://dl.profile.line-cdn.net/" + contact.pictureStatus
cl.sendVideoWithURL(msg.to, path)
except Exception as e:
raise e
print "[Command]dp executed"
elif "Getgroup image" in msg.text:
group = cl.getGroup(msg.to)
path = "http://dl.profile.line-cdn.net/" + group.pictureStatus
cl.sendImageWithURL(msg.to,path)
elif "Urlgroup image" in msg.text:
group = cl.getGroup(msg.to)
path = "http://dl.profile.line-cdn.net/" + group.pictureStatus
cl.sendText(msg.to,path)
elif "Getname" in msg.text:
key = eval(msg.contentMetadata["MENTION"])
key1 = key["MENTIONEES"][0]["M"]
contact = cl.getContact(key1)
cu = cl.channel.getCover(key1)
try:
cl.sendText(msg.to, "===[DisplayName]===\n" + contact.displayName)
except:
cl.sendText(msg.to, "===[DisplayName]===\n" + contact.displayName)
elif "Getprofile" in msg.text:
key = eval(msg.contentMetadata["MENTION"])
key1 = key["MENTIONEES"][0]["M"]
contact = cl.getContact(key1)
cu = cl.channel.getCover(key1)
path = str(cu)
image = "http://dl.profile.line-cdn.net/" + contact.pictureStatus
try:
cl.sendText(msg.to,"Nama :\n" + contact.displayName + "\n\nBio :\n" + contact.statusMessage)
cl.sendText(msg.to,"Profile Picture " + contact.displayName)
cl.sendImageWithURL(msg.to,image)
cl.sendText(msg.to,"Cover " + contact.displayName)
cl.sendImageWithURL(msg.to,path)
except:
pass
elif "Getcontact" in msg.text:
key = eval(msg.contentMetadata["MENTION"])
key1 = key["MENTIONEES"][0]["M"]
mmid = cl.getContact(key1)
msg.contentType = 13
msg.contentMetadata = {"mid": key1}
cl.sendMessage(msg)
elif "Getinfo" in msg.text:
key = eval(msg.contentMetadata["MENTION"])
key1 = key["MENTIONEES"][0]["M"]
contact = cl.getContact(key1)
cu = cl.channel.getCover(key1)
try:
cl.sendText(msg.to,"Nama :\n" + contact.displayName + "\n\nMid :\n" + contact.mid + "\n\nBio :\n" + contact.statusMessage + "\n\nProfile Picture :\nhttp://dl.profile.line-cdn.net/" + contact.pictureStatus + "\n\nHeader :\n" + str(cu))
except:
cl.sendText(msg.to,"Nama :\n" + contact.displayName + "\n\nMid :\n" + contact.mid + "\n\nBio :\n" + contact.statusMessage + "\n\nProfile Picture :\n" + str(cu))
elif "Getbio" in msg.text:
key = eval(msg.contentMetadata["MENTION"])
key1 = key["MENTIONEES"][0]["M"]
contact = cl.getContact(key1)
cu = cl.channel.getCover(key1)
try:
cl.sendText(msg.to, "===[StatusMessage]===\n" + contact.statusMessage)
except:
cl.sendText(msg.to, "===[StatusMessage]===\n" + contact.statusMessage)
elif msg.text.lower() == 'runtime':
eltime = time.time() - mulai
van = "Bot Sudah Berjalan Selama :\n"+waktu(eltime)
cl.sendText(msg.to,van)
elif "Checkdate " in msg.text:
tanggal = msg.text.replace("Checkdate ","")
r=requests.get('https://script.google.com/macros/exec?service=AKfycbw7gKzP-WYV2F5mc9RaR7yE3Ve1yN91Tjs91hp_jHSE02dSv9w&nama=ervan&tanggal='+tanggal)
data=r.text
data=json.loads(data)
lahir = data["data"]["lahir"]
usia = data["data"]["usia"]
ultah = data["data"]["ultah"]
zodiak = data["data"]["zodiak"]
cl.sendText(msg.to,"========== I N F O R M A S I ==========\n"+"Date Of Birth : "+lahir+"\nAge : "+usia+"\nUltah : "+ultah+"\nZodiak : "+zodiak+"\n========== I N F O R M A S I ==========")
elif msg.text in ["Kalender","Time","Waktu"]:
timeNow = datetime.now()
timeHours = datetime.strftime(timeNow,"(%H:%M)")
day = ["Sunday", "Monday", "Tuesday", "Wednesday", "Thursday","Friday", "Saturday"]
hari = ["Minggu", "Senin", "Selasa", "Rabu", "Kamis", "Jumat", "Sabtu"]
bulan = ["Januari", "Februari", "Maret", "April", "Mei", "Juni", "Juli", "Agustus", "September", "Oktober", "November", "Desember"]
inihari = datetime.today()
hr = inihari.strftime('%A')
bln = inihari.strftime('%m')
for i in range(len(day)):
if hr == day[i]: hasil = hari[i]
for k in range(0, len(bulan)):
if bln == str(k): bln = bulan[k-1]
rst = hasil + ", " + inihari.strftime('%d') + " - " + bln + " - " + inihari.strftime('%Y') + "\nJam : [ " + inihari.strftime('%H:%M:%S') + " ]"
cl.sendText(msg.to, rst)
elif "SearchID: " in msg.text:
userid = msg.text.replace("SearchID: ","")
contact = cl.findContactsByUserid(userid)
msg.contentType = 13
msg.contentMetadata = {'mid': contact.mid}
cl.sendMessage(msg)
elif "Searchid: " in msg.text:
userid = msg.text.replace("Searchid: ","")
contact = cl.findContactsByUserid(userid)
msg.contentType = 13
msg.contentMetadata = {'mid': contact.mid}
cl.sendMessage(msg)
elif "removechat" in msg.text.lower():
if msg.from_ in admin:
try:
cl.removeAllMessages(op.param2)
ki.removeAllMessages(op.param2)
kk.removeAllMessages(op.param2)
kc.removeAllMessages(op.param2)
kr.removeAllMessages(op.param2)
print "[Command] Remove Chat"
cl.sendText(msg.to,"Done")
except Exception as error:
print error
cl.sendText(msg.to,"Error")
elif "Invitemeto: " in msg.text:
if msg.from_ in admin:
gid = msg.text.replace("Invitemeto: ","")
if gid == "":
cl.sendText(msg.to,"Invalid group id")
else:
try:
cl.findAndAddContactsByMid(msg.from_)
ki.findAndAddContactsByMid(msg.from_)
kk.findAndAddContactsByMid(msg.from_)
kc.findAndAddContactsByMid(msg.from_)
kr.findAndAddContactsByMid(msg.from_)
random.choice(KAC).inviteIntoGroup(gid,[msg.from_])
except:
cl.sendText(msg.to,"Mungkin Saya Tidak Di Dalaam Grup Itu")
elif msg.text in ["Glist"]:
cl.sendText(msg.to, "Tunggu Sebentar. . .")
gid = cl.getGroupIdsJoined()
h = ""
for i in gid:
h += "╠➩" + "%s\n" % (cl.getGroup(i).name +" ~> ["+str(len(cl.getGroup(i).members))+"]")
cl.sendText(msg.to,"╔═════════════════════════\n║ ☆☞ LIST GROUPS☜☆\n╠═════════════════════════\n" + h + "╠═════════════════════════" + "\n║ Total Groups =" +" ["+str(len(gid))+"]\n╚═════════════════════════")
elif msg.text in ["Glistmid"]:
gruplist = kr.getGroupIdsJoined()
kontak = kr.getGroups(gruplist)
num=1
msgs="═════════List GrupMid═════════"
for ids in kontak:
msgs+="\n[%i] %s" % (num, ids.id)
num=(num+1)
msgs+="\n═════════List GrupMid═════════\n\nTotal Grup : %i" % len(kontak)
kr.sendText(msg.to, msgs)
elif "Google: " in msg.text:
a = msg.text.replace("Google: ","")
b = urllib.quote(a)
cl.sendText(msg.to,"Sedang Mencari...")
cl.sendText(msg.to, "https://www.google.com/" + b)
cl.sendText(msg.to,"Itu Dia Linknya. . .")
elif "Details group: " in msg.text:
if msg.from_ in admin:
gid = msg.text.replace("Details group: ","")
if gid in [""," "]:
cl.sendText(msg.to,"Grup id tidak valid")
else:
try:
groups = cl.getGroup(gid)
if groups.members is not None:
members = str(len(groups.members))
else:
members = "0"
if groups.invitee is not None:
pendings = str(len(groups.invitee))
else:
pendings = "0"
h = "[" + groups.name + "]\n -+GroupID : " + gid + "\n -+Members : " + members + "\n -+MembersPending : " + pendings + "\n -+Creator : " + groups.creator.displayName + "\n -+GroupPicture : http://dl.profile.line.naver.jp/" + groups.pictureStatus
cl.sendText(msg.to,h)
except Exception as error:
cl.sendText(msg.to,(error))
elif "Cancel invite: " in msg.text:
if msg.from_ in admin:
gids = msg.text.replace("Cancel invite: ","")
gid = cl.getGroup(gids)
for i in gid:
if i is not None:
try:
cl.rejectGroupInvitation(i)
except:
cl.sendText(msg.to,"Error!")
break
else:
break
if gid is not None:
cl.sendText(msg.to,"Berhasil tolak undangan dari grup " + gid.name)
else:
cl.sendText(msg.to,"Grup tidak ditemukan")
elif msg.text in ["Kapten acc invite"]:
if msg.from_ in admin:
gid = cl.getGroupIdsInvited()
_list = ""
for i in gid:
if i is not None:
gids = cl.getGroup(i)
_list += gids.name
cl.acceptGroupInvitation(i)
else:
break
if gid is not None:
cl.sendText(msg.to,"Berhasil terima semua undangan dari grup :\n" + _list)
else:
cl.sendText(msg.to,"Tidak ada grup yang tertunda saat ini")
elif msg.text in ["TC1 acc invite"]:
if msg.from_ in admin:
gid = ki.getGroupIdsInvited()
_list = ""
for i in gid:
if i is not None:
gids = ki.getGroup(i)
_list += gids.name
ki.acceptGroupInvitation(i)
else:
break
if gid is not None:
ki.sendText(msg.to,"Berhasil terima semua undangan dari grup :\n" + _list)
else:
ki.sendText(msg.to,"Tidak ada grup yang tertunda saat ini")
elif msg.text in ["TC2 acc invite"]:
if msg.from_ in admin:
gid = kk.getGroupIdsInvited()
_list = ""
for i in gid:
if i is not None:
gids = kk.getGroup(i)
_list += gids.name
kk.acceptGroupInvitation(i)
else:
break
if gid is not None:
kk.sendText(msg.to,"Berhasil terima semua undangan dari grup :\n" + _list)
else:
kk.sendText(msg.to,"Tidak ada grup yang tertunda saat ini")
elif msg.text in ["TC3 acc invite"]:
if msg.from_ in admin:
gid = kc.getGroupIdsInvited()
_list = ""
for i in gid:
if i is not None:
gids = kc.getGroup(i)
_list += gids.name
kc.acceptGroupInvitation(i)
else:
break
if gid is not None:
kc.sendText(msg.to,"Berhasil terima semua undangan dari grup :\n" + _list)
else:
kc.sendText(msg.to,"Tidak ada grup yang tertunda saat ini")
elif msg.text in ["TC4 acc invite"]:
if msg.from_ in admin:
gid = kr.getGroupIdsInvited()
_list = ""
for i in gid:
if i is not None:
gids = kr.getGroup(i)
_list += gids.name
kr.acceptGroupInvitation(i)
else:
break
if gid is not None:
kr.sendText(msg.to,"Berhasil terima semua undangan dari grup :\n" + _list)
else:
kr.sendText(msg.to,"Tidak ada grup yang tertunda saat ini")
elif "Gif gore" in msg.text:
gif = ("https://media.giphy.com/media/l2JHVsQiOZrNMGzYs/giphy.gif","https://media.giphy.com/media/OgltQ2hbilzJS/200w.gif")
gore = random.choice(gif)
cl.sendGifWithURL(msg.to,gore)
if op.type == 59:
print op
except Exception as error:
print error
while True:
try:
Ops = cl.fetchOps(cl.Poll.rev, 5)
except EOFError:
raise Exception("It might be wrong revision\n" + str(cl.Poll.rev))
for Op in Ops:
if (Op.type != OpType.END_OF_OPERATION):
cl.Poll.rev = max(cl.Poll.rev, Op.revision)
bot(Op)
| 40.951759 | 443 | 0.429363 | [
"MIT"
] | azmi155/mu | ma.py | 177,124 | Python |
# -*- coding: utf-8 -*-
import tensorflow as tf
import numpy as np
import matplotlib.pyplot as plt
import time
from PIL import Image
import random
import os
from sample import sample_conf
from tensorflow.python.framework.errors_impl import NotFoundError
# 设置以下环境变量可开启CPU识别
# os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
# os.environ["CUDA_VISIBLE_DEVICES"] = "-1"
class TrainError(Exception):
pass
class TrainModel(object):
def __init__(self, img_path, char_set, model_save_dir, verify=False):
# 模型路径
self.model_save_dir = model_save_dir
# 打乱文件顺序+校验图片格式
self.img_path = img_path
self.img_list = os.listdir(img_path)
# 校验格式
if verify:
self.confirm_image_suffix()
# 打乱文件顺序
random.seed(time.time())
random.shuffle(self.img_list)
# 获得图片宽高和字符长度基本信息
label, captcha_array = self.gen_captcha_text_image(self.img_list[0])
captcha_shape = captcha_array.shape
captcha_shape_len = len(captcha_shape)
if captcha_shape_len == 3:
image_height, image_width, channel = captcha_shape
self.channel = channel
elif captcha_shape_len == 2:
image_height, image_width = captcha_shape
else:
raise TrainError("图片转换为矩阵时出错,请检查图片格式")
# 初始化变量
# 图片尺寸
self.image_height = image_height
self.image_width = image_width
# 验证码长度(位数)
self.max_captcha = len(label)
# 验证码字符类别
self.char_set = char_set
self.char_set_len = len(char_set)
# 相关信息打印
print("-->图片尺寸: {} X {}".format(image_height, image_width))
print("-->验证码长度: {}".format(self.max_captcha))
print("-->验证码共{}类 {}".format(self.char_set_len, char_set))
print("-->使用测试集为 {}".format(img_path))
# tf初始化占位符
self.X = tf.placeholder(tf.float32, [None, image_height * image_width]) # 特征向量
self.Y = tf.placeholder(tf.float32, [None, self.max_captcha * self.char_set_len]) # 标签
self.keep_prob = tf.placeholder(tf.float32) # dropout值
self.w_alpha = 0.01
self.b_alpha = 0.1
# test model input and output
print(">>> Start model test")
batch_x, batch_y = self.get_batch(0, size=100)
print(">>> input batch images shape: {}".format(batch_x.shape))
print(">>> input batch labels shape: {}".format(batch_y.shape))
def gen_captcha_text_image(self, img_name):
"""
返回一个验证码的array形式和对应的字符串标签
:return:tuple (str, numpy.array)
"""
# 标签
label = img_name.split("_")[0]
# 文件
img_file = os.path.join(self.img_path, img_name)
captcha_image = Image.open(img_file)
captcha_array = np.array(captcha_image) # 向量化
return label, captcha_array
@staticmethod
def convert2gray(img):
"""
图片转为灰度图,如果是3通道图则计算,单通道图则直接返回
:param img:
:return:
"""
if len(img.shape) > 2:
r, g, b = img[:, :, 0], img[:, :, 1], img[:, :, 2]
gray = 0.2989 * r + 0.5870 * g + 0.1140 * b
return gray
else:
return img
def text2vec(self, text):
"""
转标签为oneHot编码
:param text: str
:return: numpy.array
"""
text_len = len(text)
if text_len > self.max_captcha:
raise ValueError('验证码最长{}个字符'.format(self.max_captcha))
vector = np.zeros(self.max_captcha * self.char_set_len)
for i, ch in enumerate(text):
idx = i * self.char_set_len + self.char_set.index(ch)
vector[idx] = 1
return vector
def get_batch(self, n, size=128):
batch_x = np.zeros([size, self.image_height * self.image_width]) # 初始化
batch_y = np.zeros([size, self.max_captcha * self.char_set_len]) # 初始化
max_batch = int(len(self.img_list) / size)
# print(max_batch)
if max_batch - 1 < 0:
raise TrainError("训练集图片数量需要大于每批次训练的图片数量")
if n > max_batch - 1:
n = n % max_batch
s = n * size
e = (n + 1) * size
this_batch = self.img_list[s:e]
# print("{}:{}".format(s, e))
for i, img_name in enumerate(this_batch):
label, image_array = self.gen_captcha_text_image(img_name)
image_array = self.convert2gray(image_array) # 灰度化图片
batch_x[i, :] = image_array.flatten() / 255 # flatten 转为一维
batch_y[i, :] = self.text2vec(label) # 生成 oneHot
return batch_x, batch_y
def confirm_image_suffix(self):
# 在训练前校验所有文件格式
print("开始校验所有图片后缀")
for index, img_name in enumerate(self.img_list):
print("{} image pass".format(index), end='\r')
if not img_name.endswith(sample_conf['image_suffix']):
raise TrainError('confirm images suffix:you request [.{}] file but get file [{}]'
.format(sample_conf['image_suffix'], img_name))
print("所有图片格式校验通过")
def model(self):
x = tf.reshape(self.X, shape=[-1, self.image_height, self.image_width, 1])
print(">>> input x: {}".format(x))
# 卷积层1
wc1 = tf.get_variable(name='wc1', shape=[3, 3, 1, 32], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
bc1 = tf.Variable(self.b_alpha * tf.random_normal([32]))
conv1 = tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(x, wc1, strides=[1, 1, 1, 1], padding='SAME'), bc1))
conv1 = tf.nn.max_pool(conv1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
conv1 = tf.nn.dropout(conv1, self.keep_prob)
# 卷积层2
wc2 = tf.get_variable(name='wc2', shape=[3, 3, 32, 64], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
bc2 = tf.Variable(self.b_alpha * tf.random_normal([64]))
conv2 = tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(conv1, wc2, strides=[1, 1, 1, 1], padding='SAME'), bc2))
conv2 = tf.nn.max_pool(conv2, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
conv2 = tf.nn.dropout(conv2, self.keep_prob)
# 卷积层3
wc3 = tf.get_variable(name='wc3', shape=[3, 3, 64, 128], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
bc3 = tf.Variable(self.b_alpha * tf.random_normal([128]))
conv3 = tf.nn.relu(tf.nn.bias_add(tf.nn.conv2d(conv2, wc3, strides=[1, 1, 1, 1], padding='SAME'), bc3))
conv3 = tf.nn.max_pool(conv3, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
conv3 = tf.nn.dropout(conv3, self.keep_prob)
print(">>> convolution 3: ", conv3.shape)
next_shape = conv3.shape[1] * conv3.shape[2] * conv3.shape[3]
# 全连接层1
wd1 = tf.get_variable(name='wd1', shape=[next_shape, 1024], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
bd1 = tf.Variable(self.b_alpha * tf.random_normal([1024]))
dense = tf.reshape(conv3, [-1, wd1.get_shape().as_list()[0]])
dense = tf.nn.relu(tf.add(tf.matmul(dense, wd1), bd1))
dense = tf.nn.dropout(dense, self.keep_prob)
# 全连接层2
wout = tf.get_variable('name', shape=[1024, self.max_captcha * self.char_set_len], dtype=tf.float32,
initializer=tf.contrib.layers.xavier_initializer())
bout = tf.Variable(self.b_alpha * tf.random_normal([self.max_captcha * self.char_set_len]))
y_predict = tf.add(tf.matmul(dense, wout), bout)
return y_predict
def train_cnn(self):
y_predict = self.model()
print(">>> input batch predict shape: {}".format(y_predict.shape))
print(">>> End model test")
# 计算概率 损失
cost = tf.reduce_mean(tf.nn.sigmoid_cross_entropy_with_logits(logits=y_predict, labels=self.Y))
# 梯度下降
optimizer = tf.train.AdamOptimizer(learning_rate=0.0001).minimize(cost)
# 计算准确率
predict = tf.reshape(y_predict, [-1, self.max_captcha, self.char_set_len]) # 预测结果
max_idx_p = tf.argmax(predict, 2) # 预测结果
max_idx_l = tf.argmax(tf.reshape(self.Y, [-1, self.max_captcha, self.char_set_len]), 2) # 标签
# 计算准确率
correct_pred = tf.equal(max_idx_p, max_idx_l)
accuracy = tf.reduce_mean(tf.reduce_min(tf.cast(correct_pred, tf.float32), axis=1))
# 模型保存对象
saver = tf.train.Saver()
with tf.Session() as sess:
init = tf.global_variables_initializer()
sess.run(init)
# 恢复模型
if os.path.exists(self.model_save_dir):
try:
saver.restore(sess, self.model_save_dir)
# 判断捕获model文件夹中没有模型文件的错误
except NotFoundError:
print("model文件夹为空,将创建新模型")
else:
pass
step = 1
for i in range(3000):
batch_x, batch_y = self.get_batch(i, size=128)
_, cost_ = sess.run([optimizer, cost], feed_dict={self.X: batch_x, self.Y: batch_y, self.keep_prob: 0.75})
if step % 10 == 0:
batch_x_test, batch_y_test = self.get_batch(i, size=100)
acc = sess.run(accuracy, feed_dict={self.X: batch_x_test, self.Y: batch_y_test, self.keep_prob: 1.})
print("第{}次训练 >>> 准确率为 {} >>> loss {}".format(step, acc, cost_))
# 准确率达到99%后保存并停止
if acc > 0.99:
saver.save(sess, self.model_save_dir)
break
# 每训练500轮就保存一次
if i % 500 == 0:
saver.save(sess, self.model_save_dir)
step += 1
saver.save(sess, self.model_save_dir)
def recognize_captcha(self):
label, captcha_array = self.gen_captcha_text_image(random.choice(self.img_list))
f = plt.figure()
ax = f.add_subplot(111)
ax.text(0.1, 0.9, "origin:" + label, ha='center', va='center', transform=ax.transAxes)
plt.imshow(captcha_array)
# 预测图片
image = self.convert2gray(captcha_array)
image = image.flatten() / 255
y_predict = self.model()
saver = tf.train.Saver()
with tf.Session() as sess:
saver.restore(sess, self.model_save_dir)
predict = tf.argmax(tf.reshape(y_predict, [-1, self.max_captcha, self.char_set_len]), 2)
text_list = sess.run(predict, feed_dict={self.X: [image], self.keep_prob: 1.})
predict_text = text_list[0].tolist()
print("正确: {} 预测: {}".format(label, predict_text))
# 显示图片和预测结果
p_text = ""
for p in predict_text:
p_text += str(self.char_set[p])
print(p_text)
plt.text(20, 1, 'predict:{}'.format(p_text))
plt.show()
def main():
train_image_dir = sample_conf["train_image_dir"]
char_set = sample_conf["char_set"]
model_save_dir = sample_conf["model_save_dir"]
tm = TrainModel(train_image_dir, char_set, model_save_dir, verify=False)
tm.train_cnn() # 开始训练模型
# tm.recognize_captcha() # 识别图片示例
if __name__ == '__main__':
main()
| 39.534965 | 122 | 0.580437 | [
"Apache-2.0"
] | shineyjg/cnn_captcha | train_model.py | 12,141 | Python |
# -*- coding: utf-8 -*-
# Licensed under a 3-clause BSD style license - see LICENSE.rst
# The idea for this module (but no code) was borrowed from the
# quantities (http://pythonhosted.org/quantities/) package.
"""Helper functions for Quantity.
In particular, this implements the logic that determines scaling and result
units for a given ufunc, given input units.
"""
from fractions import Fraction
import numpy as np
from . import UFUNC_HELPERS, UNSUPPORTED_UFUNCS
from ..core import (UnitsError, UnitConversionError, UnitTypeError,
dimensionless_unscaled, get_current_unit_registry)
def _d(unit):
if unit is None:
return dimensionless_unscaled
else:
return unit
def get_converter(from_unit, to_unit):
"""Like Unit._get_converter, except returns None if no scaling is needed,
i.e., if the inferred scale is unity."""
try:
scale = from_unit._to(to_unit)
except UnitsError:
return from_unit._apply_equivalencies(
from_unit, to_unit, get_current_unit_registry().equivalencies)
except AttributeError:
raise UnitTypeError("Unit '{0}' cannot be converted to '{1}'"
.format(from_unit, to_unit))
if scale == 1.:
return None
else:
return lambda val: scale * val
def get_converters_and_unit(f, unit1, unit2):
converters = [None, None]
# By default, we try adjusting unit2 to unit1, so that the result will
# be unit1 as well. But if there is no second unit, we have to try
# adjusting unit1 (to dimensionless, see below).
if unit2 is None:
if unit1 is None:
# No units for any input -- e.g., np.add(a1, a2, out=q)
return converters, dimensionless_unscaled
changeable = 0
# swap units.
unit2 = unit1
unit1 = None
elif unit2 is unit1:
# ensure identical units is fast ("==" is slow, so avoid that).
return converters, unit1
else:
changeable = 1
# Try to get a converter from unit2 to unit1.
if unit1 is None:
try:
converters[changeable] = get_converter(unit2,
dimensionless_unscaled)
except UnitsError:
# special case: would be OK if unitless number is zero, inf, nan
converters[1-changeable] = False
return converters, unit2
else:
return converters, dimensionless_unscaled
else:
try:
converters[changeable] = get_converter(unit2, unit1)
except UnitsError:
raise UnitConversionError(
"Can only apply '{0}' function to quantities "
"with compatible dimensions"
.format(f.__name__))
return converters, unit1
# SINGLE ARGUMENT UFUNC HELPERS
#
# The functions below take a single argument, which is the quantity upon which
# the ufunc is being used. The output of the helper function should be two
# values: a list with a single converter to be used to scale the input before
# it is being passed to the ufunc (or None if no conversion is needed), and
# the unit the output will be in.
def helper_onearg_test(f, unit):
return ([None], None)
def helper_invariant(f, unit):
return ([None], _d(unit))
def helper_square(f, unit):
return ([None], unit ** 2 if unit is not None else dimensionless_unscaled)
def helper_reciprocal(f, unit):
return ([None], unit ** -1 if unit is not None else dimensionless_unscaled)
one_half = 0.5 # faster than Fraction(1, 2)
one_third = Fraction(1, 3)
def helper_sqrt(f, unit):
return ([None], unit ** one_half if unit is not None
else dimensionless_unscaled)
def helper_cbrt(f, unit):
return ([None], (unit ** one_third if unit is not None
else dimensionless_unscaled))
def helper_modf(f, unit):
if unit is None:
return [None], (dimensionless_unscaled, dimensionless_unscaled)
try:
return ([get_converter(unit, dimensionless_unscaled)],
(dimensionless_unscaled, dimensionless_unscaled))
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"dimensionless quantities"
.format(f.__name__))
def helper__ones_like(f, unit):
return [None], dimensionless_unscaled
def helper_dimensionless_to_dimensionless(f, unit):
if unit is None:
return [None], dimensionless_unscaled
try:
return ([get_converter(unit, dimensionless_unscaled)],
dimensionless_unscaled)
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"dimensionless quantities"
.format(f.__name__))
def helper_dimensionless_to_radian(f, unit):
from ..si import radian
if unit is None:
return [None], radian
try:
return [get_converter(unit, dimensionless_unscaled)], radian
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"dimensionless quantities"
.format(f.__name__))
def helper_degree_to_radian(f, unit):
from ..si import degree, radian
try:
return [get_converter(unit, degree)], radian
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"quantities with angle units"
.format(f.__name__))
def helper_radian_to_degree(f, unit):
from ..si import degree, radian
try:
return [get_converter(unit, radian)], degree
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"quantities with angle units"
.format(f.__name__))
def helper_radian_to_dimensionless(f, unit):
from ..si import radian
try:
return [get_converter(unit, radian)], dimensionless_unscaled
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"quantities with angle units"
.format(f.__name__))
def helper_frexp(f, unit):
if not unit.is_unity():
raise UnitTypeError("Can only apply '{0}' function to "
"unscaled dimensionless quantities"
.format(f.__name__))
return [None], (None, None)
# TWO ARGUMENT UFUNC HELPERS
#
# The functions below take a two arguments. The output of the helper function
# should be two values: a tuple of two converters to be used to scale the
# inputs before being passed to the ufunc (None if no conversion is needed),
# and the unit the output will be in.
def helper_multiplication(f, unit1, unit2):
return [None, None], _d(unit1) * _d(unit2)
def helper_division(f, unit1, unit2):
return [None, None], _d(unit1) / _d(unit2)
def helper_power(f, unit1, unit2):
# TODO: find a better way to do this, currently need to signal that one
# still needs to raise power of unit1 in main code
if unit2 is None:
return [None, None], False
try:
return [None, get_converter(unit2, dimensionless_unscaled)], False
except UnitsError:
raise UnitTypeError("Can only raise something to a "
"dimensionless quantity")
def helper_ldexp(f, unit1, unit2):
if unit2 is not None:
raise TypeError("Cannot use ldexp with a quantity "
"as second argument.")
else:
return [None, None], _d(unit1)
def helper_copysign(f, unit1, unit2):
# if first arg is not a quantity, just return plain array
if unit1 is None:
return [None, None], None
else:
return [None, None], unit1
def helper_heaviside(f, unit1, unit2):
try:
converter2 = (get_converter(unit2, dimensionless_unscaled)
if unit2 is not None else None)
except UnitsError:
raise UnitTypeError("Can only apply 'heaviside' function with a "
"dimensionless second argument.")
return ([None, converter2], dimensionless_unscaled)
def helper_two_arg_dimensionless(f, unit1, unit2):
try:
converter1 = (get_converter(unit1, dimensionless_unscaled)
if unit1 is not None else None)
converter2 = (get_converter(unit2, dimensionless_unscaled)
if unit2 is not None else None)
except UnitsError:
raise UnitTypeError("Can only apply '{0}' function to "
"dimensionless quantities"
.format(f.__name__))
return ([converter1, converter2], dimensionless_unscaled)
# This used to be a separate function that just called get_converters_and_unit.
# Using it directly saves a few us; keeping the clearer name.
helper_twoarg_invariant = get_converters_and_unit
def helper_twoarg_comparison(f, unit1, unit2):
converters, _ = get_converters_and_unit(f, unit1, unit2)
return converters, None
def helper_twoarg_invtrig(f, unit1, unit2):
from ..si import radian
converters, _ = get_converters_and_unit(f, unit1, unit2)
return converters, radian
def helper_twoarg_floor_divide(f, unit1, unit2):
converters, _ = get_converters_and_unit(f, unit1, unit2)
return converters, dimensionless_unscaled
def helper_divmod(f, unit1, unit2):
converters, result_unit = get_converters_and_unit(f, unit1, unit2)
return converters, (dimensionless_unscaled, result_unit)
# list of ufuncs:
# http://docs.scipy.org/doc/numpy/reference/ufuncs.html#available-ufuncs
UNSUPPORTED_UFUNCS |= {
np.bitwise_and, np.bitwise_or, np.bitwise_xor, np.invert, np.left_shift,
np.right_shift, np.logical_and, np.logical_or, np.logical_xor,
np.logical_not}
for name in 'isnat', 'gcd', 'lcm':
# isnat was introduced in numpy 1.14, gcd+lcm in 1.15
ufunc = getattr(np, name, None)
if isinstance(ufunc, np.ufunc):
UNSUPPORTED_UFUNCS |= {ufunc}
# SINGLE ARGUMENT UFUNCS
# ufuncs that return a boolean and do not care about the unit
onearg_test_ufuncs = (np.isfinite, np.isinf, np.isnan, np.sign, np.signbit)
for ufunc in onearg_test_ufuncs:
UFUNC_HELPERS[ufunc] = helper_onearg_test
# ufuncs that return a value with the same unit as the input
invariant_ufuncs = (np.absolute, np.fabs, np.conj, np.conjugate, np.negative,
np.spacing, np.rint, np.floor, np.ceil, np.trunc,
np.positive)
for ufunc in invariant_ufuncs:
UFUNC_HELPERS[ufunc] = helper_invariant
# ufuncs that require dimensionless input and and give dimensionless output
dimensionless_to_dimensionless_ufuncs = (np.exp, np.expm1, np.exp2, np.log,
np.log10, np.log2, np.log1p)
# As found out in gh-7058, some numpy 1.13 conda installations also provide
# np.erf, even though upstream doesn't have it. We include it if present.
if isinstance(getattr(np.core.umath, 'erf', None), np.ufunc):
dimensionless_to_dimensionless_ufuncs += (np.core.umath.erf,)
for ufunc in dimensionless_to_dimensionless_ufuncs:
UFUNC_HELPERS[ufunc] = helper_dimensionless_to_dimensionless
# ufuncs that require dimensionless input and give output in radians
dimensionless_to_radian_ufuncs = (np.arccos, np.arcsin, np.arctan, np.arccosh,
np.arcsinh, np.arctanh)
for ufunc in dimensionless_to_radian_ufuncs:
UFUNC_HELPERS[ufunc] = helper_dimensionless_to_radian
# ufuncs that require input in degrees and give output in radians
degree_to_radian_ufuncs = (np.radians, np.deg2rad)
for ufunc in degree_to_radian_ufuncs:
UFUNC_HELPERS[ufunc] = helper_degree_to_radian
# ufuncs that require input in radians and give output in degrees
radian_to_degree_ufuncs = (np.degrees, np.rad2deg)
for ufunc in radian_to_degree_ufuncs:
UFUNC_HELPERS[ufunc] = helper_radian_to_degree
# ufuncs that require input in radians and give dimensionless output
radian_to_dimensionless_ufuncs = (np.cos, np.sin, np.tan, np.cosh, np.sinh,
np.tanh)
for ufunc in radian_to_dimensionless_ufuncs:
UFUNC_HELPERS[ufunc] = helper_radian_to_dimensionless
# ufuncs handled as special cases
UFUNC_HELPERS[np.sqrt] = helper_sqrt
UFUNC_HELPERS[np.square] = helper_square
UFUNC_HELPERS[np.reciprocal] = helper_reciprocal
UFUNC_HELPERS[np.cbrt] = helper_cbrt
UFUNC_HELPERS[np.core.umath._ones_like] = helper__ones_like
UFUNC_HELPERS[np.modf] = helper_modf
UFUNC_HELPERS[np.frexp] = helper_frexp
# TWO ARGUMENT UFUNCS
# two argument ufuncs that require dimensionless input and and give
# dimensionless output
two_arg_dimensionless_ufuncs = (np.logaddexp, np.logaddexp2)
for ufunc in two_arg_dimensionless_ufuncs:
UFUNC_HELPERS[ufunc] = helper_two_arg_dimensionless
# two argument ufuncs that return a value with the same unit as the input
twoarg_invariant_ufuncs = (np.add, np.subtract, np.hypot, np.maximum,
np.minimum, np.fmin, np.fmax, np.nextafter,
np.remainder, np.mod, np.fmod)
for ufunc in twoarg_invariant_ufuncs:
UFUNC_HELPERS[ufunc] = helper_twoarg_invariant
# two argument ufuncs that need compatible inputs and return a boolean
twoarg_comparison_ufuncs = (np.greater, np.greater_equal, np.less,
np.less_equal, np.not_equal, np.equal)
for ufunc in twoarg_comparison_ufuncs:
UFUNC_HELPERS[ufunc] = helper_twoarg_comparison
# two argument ufuncs that do inverse trigonometry
twoarg_invtrig_ufuncs = (np.arctan2,)
# another private function in numpy; use getattr in case it disappears
if isinstance(getattr(np.core.umath, '_arg', None), np.ufunc):
twoarg_invtrig_ufuncs += (np.core.umath._arg,)
for ufunc in twoarg_invtrig_ufuncs:
UFUNC_HELPERS[ufunc] = helper_twoarg_invtrig
# ufuncs handled as special cases
UFUNC_HELPERS[np.multiply] = helper_multiplication
UFUNC_HELPERS[np.divide] = helper_division
UFUNC_HELPERS[np.true_divide] = helper_division
UFUNC_HELPERS[np.power] = helper_power
UFUNC_HELPERS[np.ldexp] = helper_ldexp
UFUNC_HELPERS[np.copysign] = helper_copysign
UFUNC_HELPERS[np.floor_divide] = helper_twoarg_floor_divide
UFUNC_HELPERS[np.heaviside] = helper_heaviside
UFUNC_HELPERS[np.float_power] = helper_power
UFUNC_HELPERS[np.divmod] = helper_divmod
| 35.461916 | 79 | 0.677406 | [
"BSD-3-Clause"
] | PriyankaH21/astropy | astropy/units/quantity_helper/helpers.py | 14,433 | Python |
# Uses python3
import sys
def get_change(money, coins):
t = [j+1 for j in range(money+1)]
# boundary condition
t[0] = 0
for j in range(1, money+1):
for c in coins:
if c <= j:
t[j] = min(t[j], 1+t[j-c])
return t[money]
if __name__ == '__main__':
coins = [1, 3, 4]
money = int(input())
print(get_change(money, coins))
| 18.714286 | 42 | 0.516539 | [
"MIT"
] | vishweshwartyagi/Data-Structures-and-Algorithms-UCSD | 1. Algorithmic Toolbox/week5_dynamic_programming1/1_money_change_again.py | 393 | Python |
from sepal_ui import sepalwidgets as sw
from ipywidgets import dlink
from component import parameter as cp
class ParamTile(sw.Card):
def __init__(self, model):
# read the model
self.model = model
# add the base widgets
self.close = sw.Icon(children=["mdi-close"], small=True)
self.title = sw.CardTitle(
class_="pa-0 ma-0", children=[sw.Spacer(), self.close]
)
# create the widgets
self.w_target = sw.Select(
small=True,
items=[{"text": f"{i+1}0%", "value": i + 1} for i in range(cp.nb_target)],
v_model=model.target,
label="target",
dense=True,
)
self.w_weight = sw.Select(
small=True,
items=[i + 1 for i in range(cp.nb_weight)],
v_model=model.weight,
label="weight",
dense=True,
)
# link the widgets to the model
self.model.bind(self.w_target, "target").bind(self.w_weight, "weight")
# create the object
super().__init__(
max_width="500px",
class_="pa-1",
children=[self.title, self.w_target, self.w_weight],
viz=False,
disabled=False,
)
# add javascript events
self.close.on_event("click", lambda *args: self.hide())
dlink((self, "disabled"), (self, "loading"))
def reset(self):
self.w_target.v_model = None
self.w_weight.v_model = None
self.hide()
return
| 26.355932 | 86 | 0.540836 | [
"MIT"
] | 12rambau/weplan | component/tile/param_tile.py | 1,555 | Python |
#! /usr/bin/env python
"""Functions for working with the DLRN API"""
import csv
import os.path
import requests
from toolchest import yaml
from atkinson.config.manager import ConfigManager
from atkinson.logging.logger import getLogger
def _raw_fetch(url, logger):
"""
Fetch remote data and return the text output.
:param url: The URL to fetch the data from
:param logger: A logger instance to use.
:return: Raw text data, None otherwise
"""
ret_data = None
try:
req = requests.get(url)
if req.status_code == requests.codes.ok:
ret_data = req.text
except requests.exceptions.ConnectionError as error:
logger.warning(error.request)
return ret_data
def _fetch_yaml(url, logger):
"""
Fetch remote data and process the text as yaml.
:param url: The URL to fetch the data from
:param logger: A logger instance to use.
:return: Parsed yaml data in the form of a dictionary
"""
ret_data = None
raw_data = _raw_fetch(url, logger)
if raw_data is not None:
ret_data = yaml.parse(raw_data)
return ret_data
def dlrn_http_factory(host, config_file=None, link_name=None,
logger=getLogger()):
"""
Create a DlrnData instance based on a host.
:param host: A host name string to build instances
:param config_file: A dlrn config file(s) to use in addition to
the default.
:param link_name: A dlrn symlink to use. This overrides the config files
link parameter.
:param logger: An atkinson logger to use. Default is the base logger.
:return: A DlrnData instance
"""
manager = None
files = ['dlrn.yml']
if config_file is not None:
if isinstance(config_file, list):
files.extend(config_file)
else:
files.append(config_file)
local_path = os.path.realpath(os.path.dirname(__file__))
manager = ConfigManager(filenames=files, paths=local_path)
if manager is None:
return None
config = manager.config
if host not in config:
return None
link = config[host]['link']
if link_name is not None:
link = link_name
return DlrnHttpData(config[host]['url'],
config[host]['release'],
link_name=link,
logger=logger)
class DlrnHttpData():
"""A class used to interact with the dlrn API"""
def __init__(self, url, release, link_name='current', logger=getLogger()):
"""
Class constructor
:param url: The URL to the host to obtain data.
:param releases: The release name to use for lookup.
:param link_name: The name of the dlrn symlink to fetch data from.
:param logger: An atkinson logger to use. Default is the base logger.
"""
self.url = os.path.join(url, release)
self.release = release
self._logger = logger
self._link_name = link_name
self._commit_data = {}
self._fetch_commit()
def _fetch_commit(self):
"""
Fetch the commit data from dlrn
"""
full_url = os.path.join(self.url,
self._link_name,
'commit.yaml')
data = _fetch_yaml(full_url, self._logger)
if data is not None and 'commits' in data:
pkg = data['commits'][0]
if pkg['status'] == 'SUCCESS':
self._commit_data = {'name': pkg['project_name'],
'dist_hash': pkg['distro_hash'],
'commit_hash': pkg['commit_hash'],
'extended_hash': pkg.get('extended_hash')}
else:
msg = '{0} has a status of error'.format(str(pkg))
self._logger.warning(msg)
def _build_url(self):
"""
Generate a url given a commit hash and distgit hash to match the format
base/AB/CD/ABCD123_XYZ987 where ABCD123 is the commit hash and XYZ987
is a portion of the distgit hash.
:return: A string with the full URL.
"""
first = self._commit_data['commit_hash'][0:2]
second = self._commit_data['commit_hash'][2:4]
third = self._commit_data['commit_hash']
for key in ['dist_hash', 'extended_hash']:
if self._commit_data.get(key, 'None') != 'None':
third += '_' + self._commit_data[key][0:8]
return os.path.join(self.url,
first,
second,
third)
@property
def commit(self):
"""
Get the dlrn commit information
:return: A dictionary of name, dist-git hash, commit hash and
extended hash.
An empty dictionary is returned otherwise.
"""
return self._commit_data
@property
def versions(self):
"""
Get the version data for the versions.csv file and return the
data in a dictionary
:return: A dictionary of packages with commit and dist-git hashes
"""
ret_dict = {}
full_url = os.path.join(self._build_url(), 'versions.csv')
data = _raw_fetch(full_url, self._logger)
if data is not None:
data = data.replace(' ', '_')
split_data = data.split()
reader = csv.DictReader(split_data)
for row in reader:
ret_dict[row['Project']] = {'source': row['Source_Sha'],
'state': row['Status'],
'distgit': row['Dist_Sha'],
'nvr': row['Pkg_NVR']}
else:
msg = 'Could not fetch {0}'.format(full_url)
self._logger.error(msg)
return ret_dict
| 32.692308 | 79 | 0.565546 | [
"MIT"
] | jpichon/atkinson | atkinson/dlrn/http_data.py | 5,950 | Python |
"""
Module for the selection of machine learning models.
There are several different functions which can perform the model selection: all of them have an intuitive interface, but
are also powerful and flexible.
In addition, almost all these functions can optionally make plots, which sum up the performed selection in a visual way.
These different functions perform the model selection in different contexts, i.e. each function is specifically meant for a
specific scenario. Certain contexts are more specific, and other are more general.
On the whole, there are six different model selection functions, divided into two main groups:
1. functions that perform the model selection with respect to a **single dataset**;
2. functions that perform the model selection with respect to **multiple datasets**.
The six functions, sorted from the most specific context to the most general one, are:
- *hyperparameter_validation*, *hyperparameters_validation*, *models_validation* (single dataset);
- *datasets_hyperparameter_validation*, *datasets_hyperparameters_validation*, *datasets_models_validation* (multiple
datasets).
This module deeply uses the **numpy** library. It is built on the top of it. In fact, the datasets are represented as np.array.
Moreover, the plots are made using the **matplotlib** library. In addition, it is built on the top of the **sklearn** module:
- the machine learning models are represented as sklearn models (i.e. sklearn estimators);
- under the hood, the selection is performed using the grid search cross validation provided by sklearn (i.e.
GridSearchCV);
- several other operations are done using the functionalities provided by sklearn.
This module, besides the model selection functions, contains also some utilities:
- the PolynomialRegression class;
- some utility functions.
"""
import matplotlib.pyplot as plt
import numpy as np
from sklearn.utils import resample
from sklearn.model_selection import train_test_split, cross_val_score, TimeSeriesSplit, GridSearchCV
from sklearn.metrics import mean_squared_error, accuracy_score
from sklearn.preprocessing import MinMaxScaler, PolynomialFeatures
from sklearn.base import BaseEstimator
from sklearn.linear_model import LinearRegression
#----------------------------------------------------------------------------------------------------------------------------
# POLYNOMIAL REGRESSOR MODEL
class PolynomialRegression(BaseEstimator):
"""
Polynomial regression model.
It's a sklearn model: it's compliant to the sklearn estimators interface.
`Example <https://scikit-learn.org/stable/developers/develop.html>`_
Parameters
----------
degree: int
Degree to apply for the polynomial transformation.
Notes
----------
The polynomial transformation is performed using the sklearn PolynomialFeatures.
"""
def __init__(self, degree=1):
self.degree=degree
def fit(self, X, y):
self.poly_transformer = PolynomialFeatures(self.degree, include_bias=False)
self.poly_transformer.fit(X)
X = self.poly_transformer.transform(X)
self.model = LinearRegression(fit_intercept=True)
self.model.fit(X,y)
return self
def predict(self, X):
X = self.poly_transformer.transform(X)
return self.model.predict(X)
def get_params(self, deep=True):
return {"degree": self.degree}
def set_params(self, **parameters):
for parameter, value in parameters.items():
setattr(self, parameter, value)
return self
#----------------------------------------------------------------------------------------------------------------------------
# UTILITY FUNCTIONS
def compute_train_val_test(X, y, model, scale=False, test_size=0.2, time_series=False, random_state=123, n_folds=5,
regr=True):
"""
Compute the training-validation-test scores for the given model on the given dataset.
The training and test scores are simply computed by splitting the dataset into the training and test sets. The validation
score is performed applying the cross validation on the training set.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model to evaluate.
scale: bool
Indicates whether to scale or not the features in `X`.
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set.
time_series: bool
Indicates if the given dataset is a time series dataset (i.e. datasets indexed by days).
(This affects the computing of the scores).
random_state: int
Used in the training-test splitting of the dataset.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
Returns
----------
train_score: float
val_score: float
test_score: float
Notes
----------
- If `regr` is True, the returned scores are errors, computed using the MSE formula (i.e. Mean Squared Error).
Otherwise, the returned scores are accuracy measures.
- If `time_series` is False, the training-test splitting of the dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are obtained simply by splitting the dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
if regr:
scoring="neg_mean_squared_error"
else:
scoring="accuracy"
# Split into training e test.
if not time_series : # Random splitting (not time series)
X_train_80, X_test, y_train_80, y_test = train_test_split(X, y, test_size=test_size, random_state=random_state)
else: # time series splitting
train_len = int(X.shape[0]*(1-test_size))
X_train_80 = X[:train_len]
y_train_80 = y[:train_len]
X_test = X[train_len:]
y_test = y[train_len:]
if(scale): # Scale the features in X
scaler = MinMaxScaler()
scaler.fit(X_train_80)
X_train_80 = scaler.transform(X_train_80)
X_test = scaler.transform(X_test)
# Cross validation
if not time_series: # k-fold cross validation
cv = n_folds
else: # cross validation for time series
cv = TimeSeriesSplit(n_splits = n_folds)
scores = cross_val_score(model, X_train_80, y_train_80, cv=cv, scoring=scoring)
val_score = scores.mean() # validation score
if regr:
val_score = -val_score
model.fit(X_train_80,y_train_80) # Fit the model using all the training
# Compute training and test scores
train_score=0
test_score=0
if regr:
train_score = mean_squared_error(y_true=y_train_80, y_pred=model.predict(X_train_80))
test_score = mean_squared_error(y_true=y_test, y_pred=model.predict(X_test))
else:
train_score = accuracy_score(y_true=y_train_80, y_pred=model.predict(X_train_80))
test_score = accuracy_score(y_true=y_test, y_pred=model.predict(X_test))
return train_score, val_score, test_score # Return a triple
def compute_bias_variance_error(X, y, model, scale=False, N_TESTS = 20, sample_size=0.67):
"""
Compute the bias^2-variance-error scores for the given model on the given dataset.
These measures are computed in an approximate way, using `N_TESTS` random samples of size `sample_size` from the
dataset.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model to evaluate.
scale: bool
Indicates whether to scale or not the features in `X`.
(The scaling is performed using the sklearn MinMaxScaler).
N_TESTS: int
Number of samples that are made in order to compute the measures.
sample_size: float
Decimal number between 0 and 1, which indicates the proportion of the sample.
Returns
----------
bias: float
variance: float
error: float
"""
# Scale the features in `X`
if(scale):
scaler = MinMaxScaler()
scaler.fit(X)
X = scaler.transform(X)
# Vector 'vector_ypred': at the beginning is a list of lists (i.e. two dimensional list).
# In the end it will be a matrix which has as many rows as `N_TESTS` (each row corresponds to a sample) and as many
# columns as the number of instances in `X` (each column is a point of the dataset).
# Row 'i' --> there are the predictions made by the model on the sample 'i' using all the dataset points.
# Column 'j' --> there are the predictions made by the model on the point 'j' using all the `N_TESTS` samples.
vector_ypred = []
# Iterate through N_TESTS. At each iteration extract a new sample and fit the model on it.
for i in range(N_TESTS):
# Extract a new sample (sample 'i')
Xs, ys = resample(X,y, n_samples=int(sample_size*len(y)) )
# Fit the model on this sample 'i'
model.fit(Xs,ys)
# Add the predictions made by the model on all the dataset points
vector_ypred.append(list(model.predict(X)))
vector_ypred = np.array(vector_ypred) # Transform into numpy array
# Vector that has as many elements as the dataset points, and for each of them it has the associated bias^2 computed on
# the `N_TEST` samples.
vector_bias = (y - np.mean(vector_ypred, axis=0))**2
# Vector that has as many elements as the dataset points, and for each of them it has the associated variance computed on
# the `N_TEST` samples.
vector_variance = np.var(vector_ypred, axis=0)
# Vector that has as many elements as the dataset points, and for each of them it has the associated error computed on
# the `N_TEST` samples.
vector_error = np.sum((vector_ypred - y)**2, axis=0)/N_TESTS
bias = np.mean(vector_bias) # Total bias^2 of the model
variance = np.mean(vector_variance) # Total variance of the model
error = np.mean(vector_error) # Total error of the model
return bias,variance,error # Return a triple
def plot_predictions(X, y, model, scale=False, test_size=0.2, plot_type=0, xvalues=None, xlabel="Index",
title="Actual vs Predicted values", figsize=(6,6)):
"""
Plot the predictions made by the given model on the given dataset, versus its actual values.
The dataset is split into training-test sets: the former is used to train the `model`, on the latter the predictions are
made.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model used to make the predictions.
scale: bool
Indicates whether to scale or not the features in `X`.
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set.
plot_type: int
Indicates the type of the plot.
- 0 -> In the same plot two different curves are drawn: the first has on the x axis `xvalues` and on the y axis
the actual values (i.e. `y`); the second has on the x axis `xvalues` and on the y axis the computed
predicted values.
- 1 -> On the x axis the actual values are put, on the y axis the predicted ones.
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the plot.
(It's used only if `plot_type` is 0).
xlabel: str
Label of the x axis of the plot.
(It's used only if `plot_type` is 0).
title: str
Title of the plot.
figsize: tuple
Two dimensions of the plot.
Returns
----------
matplotlib.axes.Axes
The matplotlib Axes where the plot has been made.
Notes
----------
The splitting of the datasets into the training-test sets is simply made by dividing the dataset into two contiguous
sequences.
I.e. it is the same technique used usually when the dataset is a time series dataset. (This is done in order to simplify
the visualization).
For this reason, typically this function is applied on time series datasets.
"""
train_len = int(X.shape[0]*(1-test_size))
X_train_80 = X[:train_len]
y_train_80 = y[:train_len]
X_test = X[train_len:]
y_test = y[train_len:]
if(scale): # Scale the features in X
scaler = MinMaxScaler()
scaler.fit(X_train_80)
X_train_80 = scaler.transform(X_train_80)
X_test = scaler.transform(X_test)
model.fit(X_train_80,y_train_80) # Fit using all the training set
predictions = model.predict(X_test)
fig, ax = plt.subplots(figsize=figsize)
if plot_type==0:
if xvalues is None:
xvalues=range(len(X))
ax.plot(xvalues,y, 'o:', label='actual values')
ax.plot(xvalues[train_len:],predictions, 'o:', label='predicted values')
ax.legend()
elif plot_type==1:
ax.plot(y[train_len:],predictions,'o')
ax.plot([0, 1], [0, 1], 'r-',transform=ax.transAxes)
xlabel="Actual values"
ax.set_ylabel("Predicted values")
ax.set_xlabel(xlabel)
ax.set_title(title)
ax.grid()
return ax
def _plot_TrainVal_values(xvalues, train_val_scores, plot_train, xlabel, title, figsize=(6,6), bar=False):
"""
Plot the given list of training-validation scores.
This function is an auxiliary function for the model selection functions. It's meant to be private in the
module.
Parameters
----------
xvalues: list (in general iterable)
Values to put in the x axis of the plot.
train_val_scores: np.array
Two dimensional np.array, containing two columns: the first contains the trainining scores, the second the validation
scores.
Basically, it is a list of training-validation scores.
plot_train: bool
Indicates whether to plot also the training scores or to plot only the validation ones.
xlabel: str
Label of the x axis.
title: str
Title of the plot.
figsize: tuple
Two dimensions of the plot.
bar: bool
Indicates whether to plot the scores using bars or using points.
If `bar` it's True, `xvalues` must contain string (i.e. labels).
Returns
----------
matplotlib.axes.Axes
The matplotlib Axes where the plot has been made.
"""
fig, ax = plt.subplots(figsize=figsize)
if not bar: # Points
if plot_train: # Plot also the training scores
ax.plot(xvalues,train_val_scores[:,0], 'o:', label='Train')
ax.plot(xvalues,train_val_scores[:,1], 'o:', label='Validation') # Validation scores
else: # Bars
if plot_train: # Plot also the training scores
x = np.arange(len(xvalues)) # The label locations
width = 0.35 # The width of the bars
ax.bar(x-width/2,train_val_scores[:,0], width=width, label='Train')
ax.bar(x+width/2,train_val_scores[:,1], width=width, label='Validation') # Validation scores
ax.set_xticks(x)
ax.set_xticklabels(xvalues)
else:
ax.bar(xvalues,train_val_scores[:,1],label='Validation')
ax.set_xlabel(xlabel)
ax.set_title(title)
ax.grid()
ax.legend()
return ax
#----------------------------------------------------------------------------------------------------------------------------
# FUNCTIONS THAT PERFORM THE MODEL SELECTION WITH RESPECT TO A SINGLE DATASET
def hyperparameter_validation(X, y, model, hyperparameter, hyperparameter_values, scale=False, test_size=0.2,
time_series=False, random_state=123, n_folds=5, regr=True, plot=False, plot_train=False,
xvalues=None, xlabel=None, title="Hyperparameter validation", figsize=(6,6)):
"""
Select the best value for the specified hyperparameter of the specified model on the given dataset.
In other words, perform the tuning of the `hyperparameter` among the values in `hyperparameter_values`.
This selection is made using the validation score (i.e. the best hyperparameter value is the one with the best validation
score).
The validation score is computed by splitting the dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Optionally, the validation scores of the `hyperparameter_values` can be plotted, making a graphical visualization of the
selection.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model which has the specified `hyperparameter`.
hyperparameter: str
The name of the hyperparameter that has to be validated.
hyperparameter_values: list
List of values for `hyperparameter` that have to be taken into account in the selection.
scale: bool
Indicates whether to scale or not the features in `X`.
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set.
time_series: bool
Indicates if the given dataset is a time series dataset (i.e. dataset indexed by days).
(This affects the computing of the validation score).
random_state: int
Used in the training-test splitting of the dataset.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
plot: bool
Indicates whether to plot or not the validation score values.
plot_train: bool
Indicates whether to plot also the training scores.
(It's considered only if `plot` is True).
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the plot.
xlabel: str
Label of the x axis of the plot.
title: str
Title of the plot.
figsize: tuple
Two dimensions of the plot.
Returns
----------
train_val_scores: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of values in `hyperparameter_values` (i.e. number of values to be tested).
best_index: int
Index of `hyperparameter_values` that indicates which is the best hyperparameter value.
test_score: float
Test score associated with the best hyperparameter value.
ax: matplotlib.axes.Axes
The matplotlib Axes where the plot has been made.
If `plot` is False, then it is None.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
hyperparameter value is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best hyperparameter value is the one associated
with the maximum validation score.
- If `time_series` is False, the training-test splitting of the dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting the dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
param_grid = {hyperparameter:hyperparameter_values} # Create the hyperparameter grid
# Call the function for the validation of an arbitrary number of hyperparameters
params, train_val_scores, best_index, test_score = hyperparameters_validation(X, y, model, param_grid, scale=scale,
test_size=test_size,
time_series=time_series,
random_state=random_state, n_folds=n_folds,
regr=regr)
ax = None
if(plot): # Make the plot
if not xvalues: # Default values on the x axis
xvalues = hyperparameter_values
if not xlabel: # Default label on the x axis
xlabel = hyperparameter
ax = _plot_TrainVal_values(xvalues, train_val_scores, plot_train, xlabel, title, figsize)
return train_val_scores, best_index, test_score, ax
def hyperparameters_validation(X, y, model, param_grid, scale=False, test_size=0.2, time_series=False, random_state=123,
n_folds=5, regr=True):
"""
Select the best combination of values for the specified hyperparameters of the specified model on the given dataset.
In other words, perform the tuning of multiple hyperparameters.
The parameter `param_grid` is a dictionary that indicates which are the specified hyperparameters and what are the
associated values to test.
All the possible combinations of values are tested, in an exhaustive way (i.e. grid search).
This selection is made using the validation score (i.e. the best combination of hyperparameters values is the one with
the best validation score).
The validation score is computed by splitting the dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model which has the specified hyperparameters.
param_grid: dict
Dictionary which has as keys the names of the specified hyperparameters and as values the associated list of
values to test.
scale: bool
Indicates whether to scale or not the features in `X`.
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set.
time_series: bool
Indicates if the given dataset is a time series dataset (i.e. dataframe indexed by days).
(This affects the computing of the validation score).
random_state: int
Used in the training-test splitting of the dataset.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
Returns
----------
params: list
List which enumerates all the possible combinations of hyperparameters values.
It's a list of dictionaries: each dictionary represents a specific combination of hyperparameters values. (It's a
dictionary which has as keys the hyperparameters names and as values the specific associated values of that combination).
train_val_scores: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of possible combinations of the hyperparameters values.
(It has as many rows as the elements of `params`).
best_index: int
Index of `params` that indicates which is the best combination of hyperparameters values.
test_score: float
Test score associated with the best combination of hyperparameters values.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
combination of hyperparameters values is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best combination of hyperparameters values is the
one associated with the maximum validation score.
- If `time_series` is False, the training-test splitting of the dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting the dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
if regr:
scoring="neg_mean_squared_error"
else:
scoring="accuracy"
# Split into training-test sets
if not time_series : # Random splitting
X_train_80, X_test, y_train_80, y_test = train_test_split(X, y, test_size=test_size, random_state=random_state)
else: # Time series splitting
train_len = int(X.shape[0]*(1-test_size))
X_train_80 = X[:train_len]
y_train_80 = y[:train_len]
X_test = X[train_len:]
y_test = y[train_len:]
if(scale): # Scale the features in `X`
scaler = MinMaxScaler()
scaler.fit(X_train_80)
X_train_80 = scaler.transform(X_train_80)
X_test = scaler.transform(X_test)
# Cross validation strategy
if not time_series: # The strategy is the classic k-fold cross validation
cv = n_folds
else: # Time series cross validation strategy
cv = TimeSeriesSplit(n_splits = n_folds)
# Grid search
grid_search = GridSearchCV(model,param_grid,scoring=scoring,cv=cv,return_train_score=True)
grid_search.fit(X_train_80,y_train_80)
params = grid_search.cv_results_["params"] # List of all the possible combinations of hyperparameters values
# List where for all the possible combinations of hyperparameters values there is the associated training score
train_scores = grid_search.cv_results_["mean_train_score"]
# List where for all the possible combinations of hyperparameters values there is the associated validation score
val_scores = grid_search.cv_results_["mean_test_score"]
# Index of `params`, corresponding to the best combination of hyperparameters values
best_index = grid_search.best_index_
# Model with the best combination of hyperparameters values
best_model = grid_search.best_estimator_
if regr: # The scores are negative: moltiply by -1
train_scores = train_scores*(-1)
val_scores = val_scores*(-1)
train_val_scores = np.concatenate((train_scores.reshape(-1,1), val_scores.reshape(-1,1)), axis=1)
# Fit the best model on all the training set
best_model.fit(X_train_80,y_train_80)
# Compute the test score of the best model
test_score=0
if regr:
test_score = mean_squared_error(y_true=y_test, y_pred=best_model.predict(X_test))
else:
test_score = accuracy_score(y_true=y_test, y_pred=best_model.predict(X_test))
return params, train_val_scores, best_index, test_score
def models_validation(X, y, model_paramGrid_list, scale_list=None, test_size=0.2, time_series=False, random_state=123,
n_folds=5, regr=True, plot=False, plot_train=False, xvalues=None, xlabel="Models",
title="Models validation", figsize=(6,6)):
"""
Select the best model on the given dataset.
The parameter `model_paramGrid_list` is the list of the models to test. It also contains, for each model, the grid of
hyperparameters that have to be tested on that model (i.e. the grid which contains the values to test for each
specified hyperparameter of the model).
(That grid has the same structure as the `param_grid` parameter of the function `hyperparameters_validation`. See
`hyperparameters_validation`).
For each specified model, the best combination of hyperparameters values is selected in an exhaustive way (i.e. grid
search).
Actually, the function `hyperparameters_validation` is used.
(See `hyperparameters_validation`).
The selection of the best model is made using the validation score (i.e. the best model is the one with the best
validation score).
The validation score is computed by splitting the dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Optionally, the validation scores of the different models can be plotted, making a graphical visualization of the
selection.
Parameters
----------
X: np.array
Two-dimensional np.array, containing the explanatory features of the dataset.
y: np.array
Mono dimensional np.array, containing the response feature of the dataset.
model_paramGrid_list: list
List that specifies the models and the relative grids of hyperparameters to be tested.
It's a list of triples (i.e. tuples), where each triple represents a model:
- the first element is a string, which is a mnemonic name of that model;
- the second element is the sklearn model;
- the third element is the grid of hyperparameters to test for that model. It's a dictionary, with the same
structure of the parameter `param_grid` of the function `hyperparameters_validation`.
scale_list: list or bool
List of booleans, which has as many elements as the models to test (i.e. as the elements of the
`model_paramGrid_list` list).
This list indicates, for each different model, if the features in `X` have to be scaled or not.
`scale_list` can be None or False: in this case the `X` features aren't scaled for any model. `scale_list` can be
True: in this case the `X` features are scaled for all the models.
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set.
time_series: bool
Indicates if the given dataset is a time series dataset (i.e. dataset indexed by days).
(This affects the computing of the validation score).
random_state: int
Used in the training-test splitting of the dataset.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
plot: bool
Indicates whether to plot or not the validation score values.
plot_train: bool
Indicates whether to plot also the training scores.
(It's considered only if `plot` is True).
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the plot.
xlabel: str
Label of the x axis of the plot.
title: str
Title of the plot.
figsize: tuple
Two dimensions of the plot.
Returns
----------
models_train_val_score: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of models to test (i.e. number of elements in the `model_paramGrid_list` list).
models_best_params: list
List which indicates, for each model, the best combination of the hyperparameters values for that model.
It has as many elements as the models to test (i.e. as the elements of the `model_paramGrid_list` list), and it
contains dictionaries: each dictionary represents the best combination of the hyperparameters values for the
associated model.
best_index: int
Index of `model_paramGrid_list` that indicates which is the best model.
test_score: float
Test score associated with the best model.
ax: matplotlib.axes.Axes
The matplotlib Axes where the plot has been made.
If `plot` is False, then it is None.
See also
----------
hyperparameters_validation:
select the best combination of values for the specified hyperparameters of the specified model on the given dataset.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
model is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best model is the one associated with the
maximum validation score.
- If `time_series` is False, the training-test splitting of the dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting the dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
if not scale_list: # `scale_list` is either None or False
scale_list = [False]*len(model_paramGrid_list)
elif scale_list is True: # `scale_list` is True
scale_list = [True]*len(model_paramGrid_list)
# Numpy matrix (np.array) which has as many rows as the models and which has two columns, one for the training scores and
# the other for the validation scores. At the beginning it is a list of tuples.
models_train_val_score = []
# List which has as many elements as the models: for each model there is the dictionary of the best combination of
# hyperparameters values.
models_best_params = []
# List which has as many elements as the models: for each model there is the test score (associated with the best
# combination of hyperparameters values).
models_test_score = []
for i,triple in enumerate(model_paramGrid_list): # Iterate through all the cuples model-param_grid
model,param_grid = triple[1:]
# Apply the grid search on model-param_grid
params, train_val_scores, best_index, test_score = hyperparameters_validation(X, y, model, param_grid,
scale=scale_list[i],
test_size=test_size,
time_series=time_series,
random_state=random_state,
n_folds=n_folds, regr=regr)
models_train_val_score.append(tuple(train_val_scores[best_index])) # Add the row for that model
models_best_params.append(params[best_index]) # Add the element for that model
models_test_score.append(test_score) # Add the element for that model
models_train_val_score = np.array(models_train_val_score) # Transform into numpy matrix (i.e. np.array)
# Find the best index (i.e. the best model)
if regr:
best_index = np.argmin(models_train_val_score,axis=0)[1]
else:
best_index = np.argmax(models_train_val_score,axis=0)[1]
# Test score of the best model
test_score = models_test_score[best_index]
ax = None
if(plot): # Make the plot
if not xvalues: # Default values for the x axis
xvalues = [model_paramGrid_list[i][0] for i in range(len(model_paramGrid_list))]
ax = _plot_TrainVal_values(xvalues, models_train_val_score, plot_train, xlabel, title, figsize, bar=True)
return models_train_val_score, models_best_params, best_index, test_score, ax
#----------------------------------------------------------------------------------------------------------------------------
# FUNCTIONS THAT PERFORM THE MODEL SELECTION WITH RESPECT TO MULTIPLE DATASETS
def datasets_hyperparameter_validation(dataset_list, model, hyperparameter, hyperparameter_values, scale=False,
test_size=0.2, time_series=False, random_state=123, n_folds=5, regr=True, plot=False,
plot_train=False, xvalues=None, xlabel="Datasets", title="Datasets validation",
figsize=(6,6) ,verbose=False, figsize_verbose=(6,6)):
"""
Select the best dataset and the best value for the specified hyperparameter of the specified model (i.e. select the best
couple dataset-hyperparameter value).
For each dataset in `dataset_list`, all the specified values `hyperparameter_values` are tested for the specified
`hyperparameter` of `model`.
In other words, on each dataset the tuning of `hyperparameter` is performed: in fact, on each dataset, the function
`hyperparameter_validation` is applied. (See `hyperparameter_validation`).
In the end, the best couple dataset-hyperparameter value is selected.
Despite the fact that a couple dataset-hyperparameter value is selected, the main viewpoint is focused with respect to
the datasets. It's a validation focused on the datasets.
In fact, first of all, for each dataset the hyperparameter tuning is performed: in this way the best value is selected
and its relative score is associated with the dataset (i.e. it's the score of the dataset). (In other words, on each
dataset the function `hyperparameter_validation` is applied). Finally, after that, the best dataset is selected.
It's a two-levels selection.
This selection is made using the validation score (i.e. the best couple dataset-hyperparameter value is the one with the
best validation score).
The validation score is computed by splitting each dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Optionally, the validation scores of the datasets can be plotted, making a graphical visualization of the dataset
selection. This is the 'main' plot.
Moreover, still optionally, the 'secondary' plots can be done: for each dataset, the validation scores of the
`hyperparameter_values` are plotted, making a graphical visualization of the hyperparameter tuning on that dataset.
(As the plot made by the `hyperparameter_validation` function).
Parameters
----------
dataset_list: list
List of couples, where each couple is a dataset.
- The first element is X, the two-dimensional np.array containing the explanatory features of the dataset.
- The second element is y, the mono dimensional np.array containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model which has the specified `hyperparameter`.
hyperparameter: str
The name of the hyperparameter that has to be validated.
hyperparameter_values: list
List of values for `hyperparameter` that have to be taken into account in the selection.
scale: bool
Indicates whether to scale or not the features in 'X' (for all the datasets).
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set (for each dataset).
time_series: bool
Indicates if the given datasets are time series dataset (i.e. datasets indexed by days).
(This affects the computing of the validation scores).
random_state: int
Used in the training-test splitting of the datasets.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
plot: bool
Indicates whether to plot or not the validation score values of the datasets (i.e. this is the 'main' plot).
plot_train: bool
Indicates whether to plot also the training scores (both in the 'main' and 'secondary' plots).
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the 'main' plot.
xlabel: str
Label of the x axis of the 'main' plot.
title: str
Title of the 'main' plot.
figsize: tuple
Two dimensions of the 'main' plot.
verbose: bool
If True, for each dataset are plotted the validation scores of the hyperparameter tuning (these are the 'secondary'
plots).
(See 'hyperparameter_validation').
figsize_verbose: tuple
Two dimensions of the 'secondary' plots.
Returns
----------
datasets_train_val_score: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of datasets to test, i.e. as the number of elements in `dataset_list`.
datasets_best_hyperparameter_value: list
List which has as many elements as the number of the datasets (i.e. as the number of elements in `dataset_list`). For
each dataset, it contains the best `hyperparameter` value on that dataset.
best_index: int
Index of `dataset_list` that indicates which is the best dataset.
test_score: float
Test score associated with the best couple dataset-hyperparameter value.
axes: list
List of the matplotlib Axes where the plots have been made.
Firstly, the 'secondary' plots are put (if any). And, as last, the 'main' plot is put (if any).
If no plot has been made, `axes` is an empty list.
See also
----------
hyperparameter_validation:
select the best value for the specified hyperparameter of the specified model on the given dataset.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
couple dataset-hyperparameter value is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best couple is the one associated with the
maximum validation score.
- If `time_series` is False, the training-test splitting of each dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting each dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
# numpy matrix (i.e. np.array) which has as many rows as the datasets, and it has the training and validation scores as
# columns. At the beginning it is a list.
datasets_train_val_score = []
# List which contains, for each dataset, the best hyperparameter value
datasets_best_hyperparameter_value = []
# List which contains, for each dataset, its test score (associated with the best hyperparameter value)
datasets_test_score = []
# List of axes
axes = []
for i,dataset in enumerate(dataset_list): # Iterate through all the datasets
X,y = dataset
# Perform the hyperparameter tuning on the current dataset
train_val_scores, best_index, test_score, ax = hyperparameter_validation(X, y, model, hyperparameter,
hyperparameter_values, scale=scale, test_size=test_size, time_series=time_series,
random_state=random_state, n_folds=n_folds, regr=regr, plot=verbose, plot_train=plot_train,
xvalues=hyperparameter_values, xlabel=hyperparameter,
title="Dataset "+str(i)+" : hyperparameter validation", figsize=figsize_verbose)
datasets_train_val_score.append(tuple(train_val_scores[best_index,:])) # Add the row related to that dataset
datasets_best_hyperparameter_value.append(hyperparameter_values[best_index]) # Add the element related to that dataset
datasets_test_score.append(test_score) # Add the row related to that dataset
if ax:
axes.append(ax)
datasets_train_val_score = np.array(datasets_train_val_score) # Transform into numpy
# Find the best index, i.e. the best dataset (more precisely, the best couple dataset-hyperparameter value)
if regr:
best_index = np.argmin(datasets_train_val_score,axis=0)[1]
else:
best_index = np.argmax(datasets_train_val_score,axis=0)[1]
# Test score of the best couple dataset-hyperparameter value
test_score = datasets_test_score[best_index]
if(plot): # Make the plot
if not xvalues: # Default values on the x axis
xvalues = range(len(dataset_list))
ax = _plot_TrainVal_values(xvalues,datasets_train_val_score,plot_train,xlabel,title,figsize, bar=True)
axes.append(ax)
return datasets_train_val_score, datasets_best_hyperparameter_value, best_index, test_score, axes
def datasets_hyperparameters_validation(dataset_list, model, param_grid, scale=False, test_size=0.2, time_series=False,
random_state=123, n_folds=5, regr=True, plot=False, plot_train=False, xvalues=None,
xlabel="Datasets", title="Datasets validation",figsize=(6,6)):
"""
Select the best dataset and the best combination of values for the specified hyperparameters of the specified model (i.e.
select the best couple dataset-combination of hyperparameters values).
For each dataset in `dataset_list`, all the possible combinations of the hyperparameters values for `model` (specified
with `param_grid`) are tested.
In other words, on each dataset the tuning of the specified hyperparameters is performed in an exhaustive way: in fact,
on each dataset, the function `hyperparameters_validation` is applied. (See `hyperparameters_validation`).
In the end, the best couple dataset-combination of hyperparameters values is selected.
Despite the fact that a couple dataset-combination of hyperparameters values is selected, the main viewpoint is focused
with respect to the datasets. It's a validation focused on the datasets.
In fact, first of all, for each dataset the hyperparameters tuning is performed: in this way the best combination of
values is selected and its relative score is associated with the dataset (i.e. it's the score of the dataset). (In other
words, on each dataset the function `hyperparameters_validation` is applied). Finally, after that, the best dataset is
selected. It's a two-levels selection.
This selection is made using the validation score (i.e. the best couple dataset-combination of hyperparameters values, is
the one with best validation score).
The validation score is computed by splitting each dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Optionally, the validation scores of the datasets can be plotted, making a graphical visualization of the dataset
selection.
Parameters
----------
dataset_list: list
List of couple, where each couple is a dataset.
- The first element is X, the two-dimensional np.array containing the explanatory features of the dataset.
- The second element is y, the mono dimensional np.array containing the response feature of the dataset.
model: sklearn.base.BaseEstimator
Model which has the specified hyperparameters.
param_grid: dict
Dictionary which has as keys the names of the specified hyperparameters and as values the associated list of
values to test.
scale: bool
Indicates whether to scale or not the features in 'X' (for all the datasets).
(The scaling is performed using the sklearn MinMaxScaler).
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set (for each dataset).
time_series: bool
Indicates if the given datasets are time series datasets (i.e. datasets indexed by days).
(This affects the computing of the validation score).
random_state: int
Used in the training-test splitting of the datasets.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
plot: bool
Indicates whether to plot or not the validation score values of the datasets.
plot_train: bool
Indicates whether to plot also the training scores.
(It's considered only if `plot` is True).
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the plot.
xlabel: str
Label of the x axis of the plot.
title: str
Title of the plot.
figsize: tuple
Two dimensions of the plot.
Returns
----------
datasets_train_val_score: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of datasets to test, i.e. as the number of elements in `dataset_list`.
datasets_best_params: list
List which has as many elements as the number of the datasets (i.e. as the number of elements in `dataset_list`). For
each dataset, it contains the best combination of hyperparameters values on that dataset.
Each combination is represented as a dictionary, with keys the hyperparameters names and values the associated
values.
best_index: int
Index of `dataset_list` that indicates which is the best dataset.
test_score: float
Test score associated with the best couple dataset-combination of hyperparameters values.
ax: matplotlib.axes.Axes
The matplotlib Axes where the plot has been made.
See also
----------
hyperparameters_validation:
select the best combination of values for the specified hyperparameters of the specified model on the given dataset.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
couple dataset-combination of hyperparameters values is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best couple is the one associated with the
maximum validation score.
- If `time_series` is False, the training-test splitting of each dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting each dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
# numpy matrix (i.e. np.array) which has as many rows as the datasets, and it has the training and validation scores as
# columns . At the beginning it is a list.
datasets_train_val_score = []
# List which contains, for each dataset, the best combination of hyperparameters values (i.e. a dictionary)
datasets_best_params = []
# List which contains, for each dataset, its test score (associated to the best combination of hyperparameters values)
datasets_test_score = []
for X,y in dataset_list: # Iterate through all the datasets
# Perform the exaustive hyperparameters tuning on the current dataset
params, train_val_scores, best_index, test_score = hyperparameters_validation(X, y, model, param_grid, scale=scale,
test_size=test_size,
time_series=time_series,
random_state=random_state,
n_folds=n_folds, regr=regr)
datasets_train_val_score.append(tuple(train_val_scores[best_index,:])) # Add the row related to that dataset
datasets_best_params.append(params[best_index]) # Add the element related to that dataset
datasets_test_score.append(test_score) # Add the row related to that dataset
datasets_train_val_score = np.array(datasets_train_val_score) # Transform into numpy
# Find the best index, i.e. the best dataset (more precisely, the best couple dataset-combination of hyperparameters
# values)
if regr:
best_index = np.argmin(datasets_train_val_score,axis=0)[1]
else:
best_index = np.argmax(datasets_train_val_score,axis=0)[1]
# Test score of the best couple dataset-combination of hyperparameters values
test_score = datasets_test_score[best_index]
ax = None
if(plot): # Make the plot
if not xvalues: # Default values on the x axis
xvalues = range(len(dataset_list))
ax = _plot_TrainVal_values(xvalues,datasets_train_val_score,plot_train,xlabel,title,figsize, bar=True)
return datasets_train_val_score, datasets_best_params, best_index, test_score, ax
def datasets_models_validation(dataset_list, model_paramGrid_list, scale_list=None, test_size=0.2, time_series=False,
random_state=123, n_folds=5, regr=True, plot=False, plot_train=False, xvalues=None,
xlabel="Datasets", title="Datasets validation", figsize=(6,6) ,verbose=False,
figsize_verbose=(6,6)):
"""
Select the best dataset and the best model (i.e. select the best couple dataset-model).
For each dataset in `dataset_list`, all the models in `model_paramGrid_list` are tested: each model is tested performing
an exhaustive tuning of the specified hyperparameters. In fact, `model_paramGrid_list` also contains, for each model, the
grid of the hyperparameters that have to be tested on that model (i.e. the grid which contains the values to test for
each specified hyperparameter of the model).
In other words, on each dataset the selection of the best model is performed: in fact, on each dataset, the function
`models_validation` is applied. (See `models_validation`).
In the end, the best couple dataset-model is selected.
Despite the fact that a couple dataset-model is selected, the main viewpoint is focused with respect to the datasets.
It's a validation focused on the datasets.
In fact, first of all, for each dataset the model selection is performed: in this way the best model is selected
and its relative score is associated with the dataset (i.e. it's the score of the dataset). (In other words, on each
dataset the function `models_validation` is applied). Finally, after that, the best dataset is selected.
It's a two-levels selection.
This selection is made using the validation score (i.e. the best couple dataset-model is the one with best validation
score).
The validation score is computed by splitting each dataset into the training-test sets and then by applying the cross
validation on the training set.
Additionally, the training and test scores are also computed.
Optionally, the validation scores of the datasets can be plotted, making a graphical visualization of the dataset
selection. This is the 'main' plot.
Moreover, still optionally, the 'secondary' plots can be done: for each dataset, the validation scores of the models are
plotted, making a graphical visualization of the models selection on that dataset. (As the plot made by the
`models_validation` function).
Parameters
----------
dataset_list: list
List of couples, where each couple is a dataset.
- The first element is X, the two-dimensional np.array containing the explanatory features of the dataset.
- The second element is y, the mono dimensional np.array containing the response feature of the dataset.
model_paramGrid_list: list
List that specifies the models and the relative grid of hyperparameters to be tested.
It's a list of triples (i.e. tuples), where each triple represents a model:
- the first element is a string, which is a mnemonic name of that model;
- the second element is the sklearn model;
- the third element is the grid of hyperparameters to test for that model. It's a dictionary, with the same
structure of parameter `param_grid` of the function `hyperparameters_validation`.
scale_list: list or bool
List of booleans, which has as many elements as the number of models to test (i.e. number of elements in the
`model_paramGrid_list` list).
This list indicates, for each different model, if the features in 'X' have to be scaled or not (for all the datasets).
`scale_list` can be None or False: in this case the 'X' features aren't scaled for any model. `scale_list` can be
True: in this case the 'X' features are scaled for all the models.
test_size: float
Decimal number between 0 and 1, which indicates the proportion of the test set (for each dataset).
time_series: bool
Indicates if the given datasets are time series dataset (i.e. datasets indexed by days).
(This affects the computing of the validation score).
random_state: int
Used in the training-test splitting of the datasets.
n_folds: int
Indicates how many folds are made in order to compute the k-fold cross validation.
(It's used only if `time_series` is False).
regr: bool
Indicates if it's either a regression or a classification problem.
plot: bool
Indicates whether to plot or not the validation score values of the datasets (i.e. this is the 'main' plot).
plot_train: bool
Indicates whether to plot also the training scores (both in the 'main' and 'secondary' plots).
xvalues: list (in general, iterable)
Values that have to be put in the x axis of the 'main' plot.
xlabel: str
Label of the x axis of the 'main' plot.
title: str
Title of the 'main' plot.
figsize: tuple
Two dimensions of the 'main' plot.
verbose: bool
If True, for each dataset the validation scores of the models are plotted (i.e. these are the 'secondary' plots).
(See 'models_validation').
figsize_verbose: tuple
Two dimensions of the 'secondary' plots.
Returns
----------
datasets_train_val_score: np.array
Two dimensional np.array, containing two columns: the first contains the training scores, the second the validation
scores.
It has as many rows as the number of datasets to test, i.e. as the number of elements in `dataset_list`.
datasets_best_model: list
List which has as many elements as the number of the datasets (i.e. number of elements in `dataset_list`). For
each dataset, it contains the best model for that dataset.
More precisely, it is a list of triple:
- the first element is the index of `model_paramGrid_list` which indicates the best model;
- the second element is the mnemonic name of the best model;
- the third element is the best combination of hyperparameters values on that best model (i.e. it's a dictionary
which has as keys the hyperparameters names and as values their associated values).
best_index: int
Index of `dataset_list` that indicates which is the best dataset.
test_score: float
Test score associated with the best couple dataset-model.
axes: list
List of the matplotlib Axes where the plots have been made.
Firstly, the 'secondary' plots are put (if any). And, as last, the 'main' plot is put (if any).
If no plot has been made, `axes` is an empty list.
See also
----------
models_validation: select the best model on the given dataset.
Notes
----------
- If `regr` is True, the validation scores are errors (MSE, i.e. Mean Squared Errors): this means that the best
couple dataset-model is the one associated with the minimum validation score.
Otherwise, the validation scores are accuracies: this means that the best couple is the one associated with the
maximum validation score.
- If `time_series` is False, the training-test splitting of each dataset is made randomly. In addition, the cross
validation strategy performed is the classic k-fold cross validation: the number of folds is specified by `n_folds`.
Otherwise, if `time_series` is True, the training-test sets are simply obtained by splitting each dataset into two
contiguous parts. In addition, the cross validation strategy performed is the sklearn TimeSeriesSplit.
"""
# numpy matrix (i.e. np.array) which has as many rows as the datasets, and it has the training and validation scores as
# columns. At the beginning it is a list.
datasets_train_val_score = []
# List which contains, for each dataset, the best model. I.e. there is the triple index-model name-best combination of
# hyperparameters values
datasets_best_model = []
# List which contains, for each dataset, its test score (associated to the best model)
datasets_test_score = []
# List of axes
axes = []
for i,dataset in enumerate(dataset_list): # Iterate through all the datasets
X,y = dataset
# Perform the models validation on the current dataset
models_train_val_score, models_best_params, best_index, test_score, ax = models_validation(X, y,
model_paramGrid_list,
scale_list=scale_list,
test_size=test_size,
time_series=time_series,
random_state=random_state,
n_folds=n_folds,
regr=regr, plot=verbose,
plot_train=plot_train,
xlabel="Models",
title=("Dataset "+str(i)+
" : models validation"),
figsize=figsize_verbose)
datasets_train_val_score.append(tuple(models_train_val_score[best_index,:])) # Add the row related to that dataset
# Add the element related to that dataset
datasets_best_model.append((best_index,model_paramGrid_list[best_index][0],models_best_params[best_index]))
datasets_test_score.append(test_score) # Add the element related to that dataset
if ax:
axes.append(ax)
datasets_train_val_score = np.array(datasets_train_val_score) # Transform into numpy
# Find the best index, i.e. the best dataset (more precisely, the best couple dataset-model)
if regr:
best_index = np.argmin(datasets_train_val_score,axis=0)[1]
else:
best_index = np.argmax(datasets_train_val_score,axis=0)[1]
# Test score of the best couple dataset-model
test_score = datasets_test_score[best_index]
if(plot): # Make the plot
if not xvalues: # Default values on the x axis
xvalues = range(len(dataset_list))
ax = _plot_TrainVal_values(xvalues,datasets_train_val_score,plot_train,xlabel,title,figsize, bar=True)
axes.append(ax)
return datasets_train_val_score, datasets_best_model, best_index, test_score, axes
| 52.110855 | 130 | 0.664643 | [
"MIT"
] | EnricoPittini/model-selection | model_selection.py | 67,692 | Python |
from asyncio import AbstractEventLoop
from asyncio import iscoroutinefunction
from collections import defaultdict
from typing import Any
from typing import Callable
from typing import DefaultDict
from typing import Dict
from typing import NamedTuple
from typing import Set
from typing import Tuple
from .api.master_api_client import MasterApiClient
CallbackFunc = Callable[[str, Any], None]
class Callback(NamedTuple):
callback: CallbackFunc
class ParamManager:
def __init__(
self,
master_api_client: MasterApiClient,
loop: AbstractEventLoop
) -> None:
self._master_api_client = master_api_client
self._loop = loop
self._callbacks: DefaultDict[str, Set[Callback]] = defaultdict(set)
self._cache: Dict[str, Any] = {}
async def subscribe_param(
self,
key: str,
callback: CallbackFunc
) -> Tuple[Any, Callback]:
if key not in self._callbacks:
param_value = await self._master_api_client.subscribe_param(key)
self._cache[key] = param_value
else:
param_value = self._cache[key]
cb = Callback(callback)
self._callbacks[key].add(cb)
return param_value, cb
async def unsubscribe_callback(
self,
callback: Callback
) -> bool:
for key, callbacks in self._callbacks.items():
if callback in callbacks:
callbacks.discard(callback)
break
else:
return False
if not callbacks:
await self._master_api_client.unsusbcribe_param(key)
self._cache.pop(key)
self._callbacks.pop(key)
return True
def update(self, key: str, value: Any) -> bool:
self._cache[key] = value
callbacks = set()
namespace = '/'
for ns in key.split('/'):
if not ns:
continue
namespace += ns
callbacks |= set(self._callbacks.get(namespace, set()))
namespace += '/'
if not callbacks:
return False
for callback in callbacks:
if iscoroutinefunction(callback.callback):
self._loop.create_task(callback.callback(key, value))
else:
self._loop.call_soon(callback.callback, key, value)
return True
| 27.55814 | 76 | 0.613924 | [
"Apache-2.0"
] | mgrrx/aioros | src/aioros/param_manager.py | 2,370 | Python |
#!/usr/bin/env python3
import importlib.machinery as imm
import logging
import pathlib
import re
import configargparse
class ModuleInfo:
def __init__(self, path):
self.path = pathlib.Path(path)
name = str(self.path.parent / self.path.stem)
name = name.replace("/", ".")
self.name = re.sub(r"^[\.]+", "", name)
self.module = imm.SourceFileLoader(self.name, path).load_module()
if not hasattr(self.module, "get_parser"):
raise ValueError(f"{path} does not have get_parser()")
def get_parser():
parser = configargparse.ArgumentParser(
description='generate RST from argparse options',
config_file_parser_class=configargparse.YAMLConfigFileParser,
formatter_class=configargparse.ArgumentDefaultsHelpFormatter)
parser.add_argument('src', type=str, nargs='+',
help='source python files that contain get_parser() func')
return parser
# parser
args = get_parser().parse_args()
modinfo = []
for p in args.src:
if "__init__.py" in p:
continue
modinfo.append(ModuleInfo(p))
# print refs
for m in modinfo:
logging.info(f"processing: {m.path.name}")
d = m.module.get_parser().description
assert d is not None
print(f"- :ref:`{m.path.name}`: {d}")
print()
# print argparse
for m in modinfo:
cmd = m.path.name
sep = "~" * len(cmd)
print(f"""
.. _{cmd}:
{cmd}
{sep}
.. argparse::
:module: {m.name}
:func: get_parser
:prog: {cmd}
""")
| 21.927536 | 82 | 0.637145 | [
"Apache-2.0"
] | 18445864529/espnet | doc/argparse2rst.py | 1,513 | Python |
"""
util_list module. Contains the mflist class.
This classes encapsulates modflow-style list inputs away
from the individual packages. The end-user should not need to
instantiate this class directly.
some more info
"""
from __future__ import division, print_function
import os
import warnings
import numpy as np
from ..datbase import DataInterface, DataListInterface, DataType
from ..utils.recarray_utils import create_empty_recarray
try:
from numpy.lib import NumpyVersion
numpy114 = NumpyVersion(np.__version__) >= "1.14.0"
except ImportError:
numpy114 = False
class MfList(DataInterface, DataListInterface):
"""
a generic object for handling transient boundary condition lists
Parameters
----------
package : package object
The package object (of type :class:`flopy.pakbase.Package`) to which
this MfList will be added.
data : varies
the data of the transient list (optional). (the default is None)
Attributes
----------
mxact : int
the max number of active bc for any stress period
Methods
-------
add_record(kper,index,value) : None
add a record to stress period kper at index location
write_transient(f) : None
write the transient sequence to the model input file f
check_kij() : None
checks for boundaries outside of model domain - issues warnings only
See Also
--------
Notes
-----
Examples
--------
"""
def __init__(
self,
package,
data=None,
dtype=None,
model=None,
list_free_format=None,
binary=False,
):
if isinstance(data, MfList):
for attr in data.__dict__.items():
setattr(self, attr[0], attr[1])
if model is None:
self._model = package.parent
else:
self._model = model
self._package = package
return
self._package = package
if model is None:
self._model = package.parent
else:
self._model = model
if dtype is None:
assert isinstance(self.package.dtype, np.dtype)
self.__dtype = self.package.dtype
else:
self.__dtype = dtype
self.__binary = binary
self.__vtype = {}
self.__data = {}
if data is not None:
self.__cast_data(data)
self.__df = None
if list_free_format is None:
if package.parent.version == "mf2k":
list_free_format = False
self.list_free_format = list_free_format
return
@property
def name(self):
return self.package.name
@property
def mg(self):
return self._model.modelgrid
@property
def sr(self):
return self.mg.sr
@property
def model(self):
return self._model
@property
def package(self):
return self._package
@property
def data_type(self):
return DataType.transientlist
@property
def plotable(self):
return True
def get_empty(self, ncell=0):
d = create_empty_recarray(ncell, self.dtype, default_value=-1.0e10)
return d
def export(self, f, **kwargs):
from flopy import export
return export.utils.mflist_export(f, self, **kwargs)
def append(self, other):
""" append the recarrays from one MfList to another
Parameters
----------
other: variable: an item that can be cast in to an MfList
that corresponds with self
Returns
-------
dict of {kper:recarray}
"""
if not isinstance(other, MfList):
other = MfList(
self.package,
data=other,
dtype=self.dtype,
model=self._model,
list_free_format=self.list_free_format,
)
msg = (
"MfList.append(): other arg must be "
+ "MfList or dict, not {0}".format(type(other))
)
assert isinstance(other, MfList), msg
other_kpers = list(other.data.keys())
other_kpers.sort()
self_kpers = list(self.data.keys())
self_kpers.sort()
new_dict = {}
for kper in range(self._model.nper):
other_data = other[kper].copy()
self_data = self[kper].copy()
other_len = other_data.shape[0]
self_len = self_data.shape[0]
if (other_len == 0 and self_len == 0) or (
kper not in self_kpers and kper not in other_kpers
):
continue
elif self_len == 0:
new_dict[kper] = other_data
elif other_len == 0:
new_dict[kper] = self_data
else:
new_len = other_data.shape[0] + self_data.shape[0]
new_data = np.recarray(new_len, dtype=self.dtype)
new_data[:self_len] = self_data
new_data[self_len : self_len + other_len] = other_data
new_dict[kper] = new_data
return new_dict
def drop(self, fields):
"""drop fields from an MfList
Parameters
----------
fields : list or set of field names to drop
Returns
-------
dropped : MfList without the dropped fields
"""
if not isinstance(fields, list):
fields = [fields]
names = [n for n in self.dtype.names if n not in fields]
dtype = np.dtype(
[(k, d) for k, d in self.dtype.descr if k not in fields]
)
spd = {}
for k, v in self.data.items():
# because np 1.9 doesn't support indexing by list of columns
newarr = np.array([self.data[k][n] for n in names]).transpose()
newarr = np.array(list(map(tuple, newarr)), dtype=dtype).view(
np.recarray
)
for n in dtype.names:
newarr[n] = self.data[k][n]
spd[k] = newarr
return MfList(self.package, spd, dtype=dtype)
@property
def data(self):
return self.__data
@property
def df(self):
if self.__df is None:
self.__df = self.get_dataframe()
return self.__df
@property
def vtype(self):
return self.__vtype
@property
def dtype(self):
return self.__dtype
# Get the itmp for a given kper
def get_itmp(self, kper):
if kper not in list(self.__data.keys()):
return None
if self.__vtype[kper] is None:
return -1
# If an external file, have to load it
if self.__vtype[kper] == str:
return self.__fromfile(self.__data[kper]).shape[0]
if self.__vtype[kper] == np.recarray:
return self.__data[kper].shape[0]
# If not any of the above, it must be an int
return self.__data[kper]
@property
def mxact(self):
mxact = 0
for kper in list(self.__data.keys()):
mxact = max(mxact, self.get_itmp(kper))
return mxact
@property
def fmt_string(self):
"""Returns a C-style fmt string for numpy savetxt that corresponds to
the dtype"""
if self.list_free_format is not None:
use_free = self.list_free_format
else:
use_free = True
if self.package.parent.has_package("bas6"):
use_free = self.package.parent.bas6.ifrefm
# mt3d list data is fixed format
if "mt3d" in self.package.parent.version.lower():
use_free = False
fmts = []
for field in self.dtype.descr:
vtype = field[1][1].lower()
if vtype in ("i", "b"):
if use_free:
fmts.append("%9d")
else:
fmts.append("%10d")
elif vtype == "f":
if use_free:
if numpy114:
# Use numpy's floating-point formatter (Dragon4)
fmts.append("%15s")
else:
fmts.append("%15.7E")
else:
fmts.append("%10G")
elif vtype == "o":
if use_free:
fmts.append("%9s")
else:
fmts.append("%10s")
elif vtype == "s":
msg = (
"MfList.fmt_string error: 'str' type found in dtype. "
"This gives unpredictable results when "
"recarray to file - change to 'object' type"
)
raise TypeError(msg)
else:
raise TypeError(
"MfList.fmt_string error: unknown vtype in "
"field: {}".format(field)
)
if use_free:
fmt_string = " " + " ".join(fmts)
else:
fmt_string = "".join(fmts)
return fmt_string
# Private method to cast the data argument
# Should only be called by the constructor
def __cast_data(self, data):
# If data is a list, then all we can do is try to cast it to
# an ndarray, then cast again to a recarray
if isinstance(data, list):
# warnings.warn("MfList casting list to array")
try:
data = np.array(data)
except Exception as e:
raise Exception(
"MfList error: casting list to ndarray: " + str(e)
)
# If data is a dict, the we have to assume it is keyed on kper
if isinstance(data, dict):
if not list(data.keys()):
raise Exception("MfList error: data dict is empty")
for kper, d in data.items():
try:
kper = int(kper)
except Exception as e:
raise Exception(
"MfList error: data dict key "
+ "{0:s} not integer: ".format(kper)
+ str(type(kper))
+ "\n"
+ str(e)
)
# Same as before, just try...
if isinstance(d, list):
# warnings.warn("MfList: casting list to array at " +\
# "kper {0:d}".format(kper))
try:
d = np.array(d)
except Exception as e:
raise Exception(
"MfList error: casting list "
+ "to ndarray: "
+ str(e)
)
# super hack - sick of recarrays already
# if (isinstance(d,np.ndarray) and len(d.dtype.fields) > 1):
# d = d.view(np.recarray)
if isinstance(d, np.recarray):
self.__cast_recarray(kper, d)
elif isinstance(d, np.ndarray):
self.__cast_ndarray(kper, d)
elif isinstance(d, int):
self.__cast_int(kper, d)
elif isinstance(d, str):
self.__cast_str(kper, d)
elif d is None:
self.__data[kper] = -1
self.__vtype[kper] = None
else:
raise Exception(
"MfList error: unsupported data type: "
+ str(type(d))
+ " at kper "
+ "{0:d}".format(kper)
)
# A single recarray - same MfList for all stress periods
elif isinstance(data, np.recarray):
self.__cast_recarray(0, data)
# A single ndarray
elif isinstance(data, np.ndarray):
self.__cast_ndarray(0, data)
# A single filename
elif isinstance(data, str):
self.__cast_str(0, data)
else:
raise Exception(
"MfList error: unsupported data type: " + str(type(data))
)
def __cast_str(self, kper, d):
# If d is a string, assume it is a filename and check that it exists
assert os.path.exists(d), (
"MfList error: dict filename (string) '"
+ d
+ "' value for "
+ "kper {0:d} not found".format(kper)
)
self.__data[kper] = d
self.__vtype[kper] = str
def __cast_int(self, kper, d):
# If d is an integer, then it must be 0 or -1
if d > 0:
raise Exception(
"MfList error: dict integer value for "
"kper {0:10d} must be 0 or -1, "
"not {1:10d}".format(kper, d)
)
if d == 0:
self.__data[kper] = 0
self.__vtype[kper] = None
else:
self.__data[kper] = -1
self.__vtype[kper] = None
def __cast_recarray(self, kper, d):
assert d.dtype == self.__dtype, (
"MfList error: recarray dtype: "
+ str(d.dtype)
+ " doesn't match "
+ "self dtype: "
+ str(self.dtype)
)
self.__data[kper] = d
self.__vtype[kper] = np.recarray
def __cast_ndarray(self, kper, d):
d = np.atleast_2d(d)
if d.dtype != self.__dtype:
assert d.shape[1] == len(self.dtype), (
"MfList error: ndarray "
+ "shape "
+ str(d.shape)
+ " doesn't match dtype "
+ "len: "
+ str(len(self.dtype))
)
# warnings.warn("MfList: ndarray dtype does not match self " +\
# "dtype, trying to cast")
try:
self.__data[kper] = np.core.records.fromarrays(
d.transpose(), dtype=self.dtype
)
except Exception as e:
raise Exception(
"MfList error: casting ndarray to recarray: " + str(e)
)
self.__vtype[kper] = np.recarray
def get_dataframe(self, squeeze=True):
"""
Cast recarrays for stress periods into single
dataframe containing all stress periods.
Parameters
----------
squeeze : bool
Reduce number of columns in dataframe to only include
stress periods where a variable changes.
Returns
-------
df : dataframe
Dataframe of shape nrow = ncells, ncol = nvar x nper. If
the squeeze option is chosen, nper is the number of
stress periods where at least one cells is different,
otherwise it is equal to the number of keys in MfList.data.
Notes
-----
Requires pandas.
"""
try:
import pandas as pd
except Exception as e:
msg = "MfList.get_dataframe() requires pandas"
raise ImportError(msg)
# make a dataframe of all data for all stress periods
names = ["k", "i", "j"]
if "MNW2" in self.package.name:
names += ["wellid"]
# find relevant variable names
# may have to iterate over the first stress period
for per in range(self._model.nper):
if hasattr(self.data[per], "dtype"):
varnames = list(
[n for n in self.data[per].dtype.names if n not in names]
)
break
# create list of dataframes for each stress period
# each with index of k, i, j
dfs = []
for per in self.data.keys():
recs = self.data[per]
if recs is None or len(recs) == 0:
# add an empty dataframe if a stress period is
# empty (e.g. no pumping during a predevelopment
# period)
columns = names + list(
["{}{}".format(c, per) for c in varnames]
)
dfi = pd.DataFrame(data=None, columns=columns)
dfi = dfi.set_index(names)
else:
dfi = pd.DataFrame.from_records(recs)
dfg = dfi.groupby(names)
count = dfg[varnames[0]].count().rename("n")
if (count > 1).values.any():
print(
"Duplicated list entry locations aggregated "
"for kper {}".format(per)
)
for kij in count[count > 1].index.values:
print(" (k,i,j) {}".format(kij))
dfi = dfg.sum() # aggregate
dfi.columns = list(["{}{}".format(c, per) for c in varnames])
dfs.append(dfi)
df = pd.concat(dfs, axis=1)
if squeeze:
keep = []
for var in varnames:
diffcols = list([n for n in df.columns if var in n])
diff = df[diffcols].fillna(0).diff(axis=1)
diff[
"{}0".format(var)
] = 1 # always return the first stress period
changed = diff.sum(axis=0) != 0
keep.append(df.loc[:, changed.index[changed]])
df = pd.concat(keep, axis=1)
df = df.reset_index()
df.insert(len(names), "node", df.i * self._model.ncol + df.j)
return df
def add_record(self, kper, index, values):
# Add a record to possible already set list for a given kper
# index is a list of k,i,j or nodes.
# values is a list of floats.
# The length of index + values must be equal to the number of names
# in dtype
assert len(index) + len(values) == len(self.dtype), (
"MfList.add_record() error: length of index arg +"
+ "length of value arg != length of self dtype"
)
# If we already have something for this kper, then add to it
if kper in list(self.__data.keys()):
if self.vtype[kper] == int:
# If a 0 or -1, reset
self.__data[kper] = self.get_empty(1)
self.__vtype[kper] = np.recarray
elif self.vtype[kper] == str:
# If filename, load into recarray
d = self.__fromfile(self.data[kper])
d.resize(d.shape[0], d.shape[1])
self.__data[kper] = d
self.__vtype[kper] = np.recarray
elif self.vtype[kper] == np.recarray:
# Extend the recarray
self.__data[kper] = np.append(
self.__data[kper], self.get_empty(1)
)
else:
self.__data[kper] = self.get_empty(1)
self.__vtype[kper] = np.recarray
rec = list(index)
rec.extend(list(values))
try:
self.__data[kper][-1] = tuple(rec)
except Exception as e:
raise Exception(
"MfList.add_record() error: adding record to "
+ "recarray: "
+ str(e)
)
def __getitem__(self, kper):
# Get the recarray for a given kper
# If the data entry for kper is a string,
# return the corresponding recarray,
# but don't reset the value in the data dict
# assert kper in list(self.data.keys()), "MfList.__getitem__() kper " + \
# str(kper) + " not in data.keys()"
try:
kper = int(kper)
except Exception as e:
raise Exception(
"MfList error: _getitem__() passed invalid kper index:"
+ str(kper)
)
if kper not in list(self.data.keys()):
if kper == 0:
return self.get_empty()
else:
return self.data[self.__find_last_kper(kper)]
if self.vtype[kper] == int:
if self.data[kper] == 0:
return self.get_empty()
else:
return self.data[self.__find_last_kper(kper)]
if self.vtype[kper] == str:
return self.__fromfile(self.data[kper])
if self.vtype[kper] == np.recarray:
return self.data[kper]
def __setitem__(self, kper, data):
if kper in list(self.__data.keys()):
if self._model.verbose:
print("removing existing data for kper={}".format(kper))
self.data.pop(kper)
# If data is a list, then all we can do is try to cast it to
# an ndarray, then cast again to a recarray
if isinstance(data, list):
# warnings.warn("MfList casting list to array")
try:
data = np.array(data)
except Exception as e:
raise Exception(
"MfList error: casting list to ndarray: " + str(e)
)
# cast data
if isinstance(data, int):
self.__cast_int(kper, data)
elif isinstance(data, np.recarray):
self.__cast_recarray(kper, data)
# A single ndarray
elif isinstance(data, np.ndarray):
self.__cast_ndarray(kper, data)
# A single filename
elif isinstance(data, str):
self.__cast_str(kper, data)
else:
raise Exception(
"MfList error: unsupported data type: " + str(type(data))
)
# raise NotImplementedError("MfList.__setitem__() not implemented")
def __fromfile(self, f):
# d = np.fromfile(f,dtype=self.dtype,count=count)
try:
d = np.genfromtxt(f, dtype=self.dtype)
except Exception as e:
raise Exception(
"MfList.__fromfile() error reading recarray "
+ "from file "
+ str(e)
)
return d
def get_filenames(self):
kpers = list(self.data.keys())
kpers.sort()
filenames = []
first = kpers[0]
for kper in list(range(0, max(self._model.nper, max(kpers) + 1))):
# Fill missing early kpers with 0
if kper < first:
itmp = 0
kper_vtype = int
elif kper in kpers:
kper_vtype = self.__vtype[kper]
if (
self._model.array_free_format
and self._model.external_path is not None
):
# py_filepath = ''
# py_filepath = os.path.join(py_filepath,
# self._model.external_path)
filename = self.package.name[0] + "_{0:04d}.dat".format(kper)
filenames.append(filename)
return filenames
def get_filename(self, kper):
ext = "dat"
if self.binary:
ext = "bin"
return self.package.name[0] + "_{0:04d}.{1}".format(kper, ext)
@property
def binary(self):
return bool(self.__binary)
def write_transient(self, f, single_per=None, forceInternal=False):
# forceInternal overrides isExternal (set below) for cases where
# external arrays are not supported (oh hello MNW1!)
# write the transient sequence described by the data dict
nr, nc, nl, nper = self._model.get_nrow_ncol_nlay_nper()
assert hasattr(f, "read"), (
"MfList.write() error: " + "f argument must be a file handle"
)
kpers = list(self.data.keys())
kpers.sort()
first = kpers[0]
if single_per is None:
loop_over_kpers = list(range(0, max(nper, max(kpers) + 1)))
else:
if not isinstance(single_per, list):
single_per = [single_per]
loop_over_kpers = single_per
for kper in loop_over_kpers:
# Fill missing early kpers with 0
if kper < first:
itmp = 0
kper_vtype = int
elif kper in kpers:
kper_data = self.__data[kper]
kper_vtype = self.__vtype[kper]
if kper_vtype == str:
if not self._model.array_free_format:
kper_data = self.__fromfile(kper_data)
kper_vtype = np.recarray
itmp = self.get_itmp(kper)
if kper_vtype == np.recarray:
itmp = kper_data.shape[0]
elif (kper_vtype == int) or (kper_vtype is None):
itmp = kper_data
# Fill late missing kpers with -1
else:
itmp = -1
kper_vtype = int
f.write(
" {0:9d} {1:9d} # stress period {2:d}\n".format(
itmp, 0, kper + 1
)
)
isExternal = False
if (
self._model.array_free_format
and self._model.external_path is not None
and forceInternal is False
):
isExternal = True
if self.__binary:
isExternal = True
if isExternal:
if kper_vtype == np.recarray:
py_filepath = ""
if self._model.model_ws is not None:
py_filepath = self._model.model_ws
if self._model.external_path is not None:
py_filepath = os.path.join(
py_filepath, self._model.external_path
)
filename = self.get_filename(kper)
py_filepath = os.path.join(py_filepath, filename)
model_filepath = filename
if self._model.external_path is not None:
model_filepath = os.path.join(
self._model.external_path, filename
)
self.__tofile(py_filepath, kper_data)
kper_vtype = str
kper_data = model_filepath
if kper_vtype == np.recarray:
name = f.name
if self.__binary or not numpy114:
f.close()
# switch file append mode to binary
with open(name, "ab+") as f:
self.__tofile(f, kper_data)
# continue back to non-binary
f = open(name, "a")
else:
self.__tofile(f, kper_data)
elif kper_vtype == str:
f.write(" open/close " + kper_data)
if self.__binary:
f.write(" (BINARY)")
f.write("\n")
def __tofile(self, f, data):
# Write the recarray (data) to the file (or file handle) f
assert isinstance(data, np.recarray), (
"MfList.__tofile() data arg " + "not a recarray"
)
# Add one to the kij indices
lnames = [name.lower() for name in self.dtype.names]
# --make copy of data for multiple calls
d = data.copy()
for idx in ["k", "i", "j", "node"]:
if idx in lnames:
d[idx] += 1
if self.__binary:
dtype2 = []
for name in self.dtype.names:
dtype2.append((name, np.float32))
dtype2 = np.dtype(dtype2)
d = np.array(d, dtype=dtype2)
d.tofile(f)
else:
np.savetxt(f, d, fmt=self.fmt_string, delimiter="")
def check_kij(self):
names = self.dtype.names
if ("k" not in names) or ("i" not in names) or ("j" not in names):
warnings.warn(
"MfList.check_kij(): index fieldnames 'k,i,j' "
+ "not found in self.dtype names: "
+ str(names)
)
return
nr, nc, nl, nper = self._model.get_nrow_ncol_nlay_nper()
if nl == 0:
warnings.warn(
"MfList.check_kij(): unable to get dis info from " + "model"
)
return
for kper in list(self.data.keys()):
out_idx = []
data = self[kper]
if data is not None:
k = data["k"]
k_idx = np.where(np.logical_or(k < 0, k >= nl))
if k_idx[0].shape[0] > 0:
out_idx.extend(list(k_idx[0]))
i = data["i"]
i_idx = np.where(np.logical_or(i < 0, i >= nr))
if i_idx[0].shape[0] > 0:
out_idx.extend(list(i_idx[0]))
j = data["j"]
j_idx = np.where(np.logical_or(j < 0, j >= nc))
if j_idx[0].shape[0]:
out_idx.extend(list(j_idx[0]))
if len(out_idx) > 0:
warn_str = (
"MfList.check_kij(): warning the following "
+ "indices are out of bounds in kper "
+ str(kper)
+ ":\n"
)
for idx in out_idx:
d = data[idx]
warn_str += " {0:9d} {1:9d} {2:9d}\n".format(
d["k"] + 1, d["i"] + 1, d["j"] + 1
)
warnings.warn(warn_str)
def __find_last_kper(self, kper):
kpers = list(self.data.keys())
kpers.sort()
last = 0
for kkper in kpers[::-1]:
# if this entry is valid
if self.vtype[kkper] != int or self.data[kkper] != -1:
last = kkper
if kkper <= kper:
break
return kkper
def get_indices(self):
"""
a helper function for plotting - get all unique indices
"""
names = self.dtype.names
lnames = []
[lnames.append(name.lower()) for name in names]
if "k" not in lnames or "j" not in lnames:
raise NotImplementedError("MfList.get_indices requires kij")
kpers = list(self.data.keys())
kpers.sort()
indices = []
for i, kper in enumerate(kpers):
kper_vtype = self.__vtype[kper]
if (kper_vtype != int) or (kper_vtype is not None):
d = self.data[kper]
if not indices:
indices = list(zip(d["k"], d["i"], d["j"]))
else:
new_indices = list(zip(d["k"], d["i"], d["j"]))
for ni in new_indices:
if ni not in indices:
indices.append(ni)
return indices
def attribute_by_kper(self, attr, function=np.mean, idx_val=None):
assert attr in self.dtype.names
if idx_val is not None:
assert idx_val[0] in self.dtype.names
kpers = list(self.data.keys())
kpers.sort()
values = []
for kper in range(0, max(self._model.nper, max(kpers))):
if kper < min(kpers):
values.append(0)
elif kper > max(kpers) or kper not in kpers:
values.append(values[-1])
else:
kper_data = self.__data[kper]
if idx_val is not None:
kper_data = kper_data[
np.where(kper_data[idx_val[0]] == idx_val[1])
]
# kper_vtype = self.__vtype[kper]
v = function(kper_data[attr])
values.append(v)
return values
def plot(
self,
key=None,
names=None,
kper=0,
filename_base=None,
file_extension=None,
mflay=None,
**kwargs
):
"""
Plot stress period boundary condition (MfList) data for a specified
stress period
Parameters
----------
key : str
MfList dictionary key. (default is None)
names : list
List of names for figure titles. (default is None)
kper : int
MODFLOW zero-based stress period number to return. (default is zero)
filename_base : str
Base file name that will be used to automatically generate file
names for output image files. Plots will be exported as image
files if file_name_base is not None. (default is None)
file_extension : str
Valid matplotlib.pyplot file extension for savefig(). Only used
if filename_base is not None. (default is 'png')
mflay : int
MODFLOW zero-based layer number to return. If None, then all
all layers will be included. (default is None)
**kwargs : dict
axes : list of matplotlib.pyplot.axis
List of matplotlib.pyplot.axis that will be used to plot
data for each layer. If axes=None axes will be generated.
(default is None)
pcolor : bool
Boolean used to determine if matplotlib.pyplot.pcolormesh
plot will be plotted. (default is True)
colorbar : bool
Boolean used to determine if a color bar will be added to
the matplotlib.pyplot.pcolormesh. Only used if pcolor=True.
(default is False)
inactive : bool
Boolean used to determine if a black overlay in inactive
cells in a layer will be displayed. (default is True)
contour : bool
Boolean used to determine if matplotlib.pyplot.contour
plot will be plotted. (default is False)
clabel : bool
Boolean used to determine if matplotlib.pyplot.clabel
will be plotted. Only used if contour=True. (default is False)
grid : bool
Boolean used to determine if the model grid will be plotted
on the figure. (default is False)
masked_values : list
List of unique values to be excluded from the plot.
Returns
----------
out : list
Empty list is returned if filename_base is not None. Otherwise
a list of matplotlib.pyplot.axis is returned.
See Also
--------
Notes
-----
Examples
--------
>>> import flopy
>>> ml = flopy.modflow.Modflow.load('test.nam')
>>> ml.wel.stress_period_data.plot(ml.wel, kper=1)
"""
from flopy.plot import PlotUtilities
axes = PlotUtilities._plot_mflist_helper(
self,
key=key,
names=names,
kper=kper,
filename_base=filename_base,
file_extension=file_extension,
mflay=mflay,
**kwargs
)
return axes
def to_shapefile(self, filename, kper=None):
"""
Export stress period boundary condition (MfList) data for a specified
stress period
Parameters
----------
filename : str
Shapefile name to write
kper : int
MODFLOW zero-based stress period number to return. (default is None)
Returns
----------
None
See Also
--------
Notes
-----
Examples
--------
>>> import flopy
>>> ml = flopy.modflow.Modflow.load('test.nam')
>>> ml.wel.to_shapefile('test_hk.shp', kper=1)
"""
import warnings
warnings.warn(
"Deprecation warning: to_shapefile() is deprecated. use .export()"
)
# if self.sr is None:
# raise Exception("MfList.to_shapefile: SpatialReference not set")
# import flopy.utils.flopy_io as fio
# if kper is None:
# keys = self.data.keys()
# keys.sort()
# else:
# keys = [kper]
# array_dict = {}
# for kk in keys:
# arrays = self.to_array(kk)
# for name, array in arrays.items():
# for k in range(array.shape[0]):
# #aname = name+"{0:03d}_{1:02d}".format(kk, k)
# n = fio.shape_attr_name(name, length=4)
# aname = "{}{:03d}{:03d}".format(n, k+1, int(kk)+1)
# array_dict[aname] = array[k]
# fio.write_grid_shapefile(filename, self.sr, array_dict)
self.export(filename, kper=kper)
def to_array(self, kper=0, mask=False):
"""
Convert stress period boundary condition (MfList) data for a
specified stress period to a 3-D numpy array
Parameters
----------
kper : int
MODFLOW zero-based stress period number to return. (default is zero)
mask : boolean
return array with np.NaN instead of zero
Returns
----------
out : dict of numpy.ndarrays
Dictionary of 3-D numpy arrays containing the stress period data for
a selected stress period. The dictionary keys are the MfList dtype
names for the stress period data ('cond', 'flux', 'bhead', etc.).
See Also
--------
Notes
-----
Examples
--------
>>> import flopy
>>> ml = flopy.modflow.Modflow.load('test.nam')
>>> v = ml.wel.stress_period_data.to_array(kper=1)
"""
i0 = 3
unstructured = False
if "inode" in self.dtype.names:
raise NotImplementedError()
if "node" in self.dtype.names:
if "i" not in self.dtype.names and "j" not in self.dtype.names:
i0 = 1
unstructured = True
arrays = {}
for name in self.dtype.names[i0:]:
if not self.dtype.fields[name][0] == object:
if unstructured:
arr = np.zeros((self._model.nlay * self._model.ncpl,))
else:
arr = np.zeros(
(self._model.nlay, self._model.nrow, self._model.ncol)
)
arrays[name] = arr.copy()
# if this kper is not found
if kper not in self.data.keys():
kpers = list(self.data.keys())
kpers.sort()
# if this kper is before the first entry,
# (maybe) mask and return
if kper < kpers[0]:
if mask:
for name, arr in arrays.items():
arrays[name][:] = np.NaN
return arrays
# find the last kper
else:
kper = self.__find_last_kper(kper)
sarr = self.data[kper]
if np.isscalar(sarr):
# if there are no entries for this kper
if sarr == 0:
if mask:
for name, arr in arrays.items():
arrays[name][:] = np.NaN
return arrays
else:
raise Exception("MfList: something bad happened")
for name, arr in arrays.items():
if unstructured:
cnt = np.zeros(
(self._model.nlay * self._model.ncpl,), dtype=np.float
)
else:
cnt = np.zeros(
(self._model.nlay, self._model.nrow, self._model.ncol),
dtype=np.float,
)
# print(name,kper)
for rec in sarr:
if unstructured:
arr[rec["node"]] += rec[name]
cnt[rec["node"]] += 1.0
else:
arr[rec["k"], rec["i"], rec["j"]] += rec[name]
cnt[rec["k"], rec["i"], rec["j"]] += 1.0
# average keys that should not be added
if name not in ("cond", "flux"):
idx = cnt > 0.0
arr[idx] /= cnt[idx]
if mask:
arr = np.ma.masked_where(cnt == 0.0, arr)
arr[cnt == 0.0] = np.NaN
arrays[name] = arr.copy()
# elif mask:
# for name, arr in arrays.items():
# arrays[name][:] = np.NaN
return arrays
@property
def masked_4D_arrays(self):
# get the first kper
arrays = self.to_array(kper=0, mask=True)
# initialize these big arrays
m4ds = {}
for name, array in arrays.items():
m4d = np.zeros(
(
self._model.nper,
self._model.nlay,
self._model.nrow,
self._model.ncol,
)
)
m4d[0, :, :, :] = array
m4ds[name] = m4d
for kper in range(1, self._model.nper):
arrays = self.to_array(kper=kper, mask=True)
for name, array in arrays.items():
m4ds[name][kper, :, :, :] = array
return m4ds
def masked_4D_arrays_itr(self):
# get the first kper
arrays = self.to_array(kper=0, mask=True)
# initialize these big arrays
for name, array in arrays.items():
m4d = np.zeros(
(
self._model.nper,
self._model.nlay,
self._model.nrow,
self._model.ncol,
)
)
m4d[0, :, :, :] = array
for kper in range(1, self._model.nper):
arrays = self.to_array(kper=kper, mask=True)
for tname, array in arrays.items():
if tname == name:
m4d[kper, :, :, :] = array
yield name, m4d
@property
def array(self):
return self.masked_4D_arrays
@classmethod
def from_4d(cls, model, pak_name, m4ds):
"""construct an MfList instance from a dict of
(attribute_name,masked 4D ndarray
Parameters
----------
model : mbase derived type
pak_name : str package name (e.g GHB)
m4ds : {attribute name:4d masked numpy.ndarray}
Returns
-------
MfList instance
"""
sp_data = MfList.masked4D_arrays_to_stress_period_data(
model.get_package(pak_name).get_default_dtype(), m4ds
)
return cls(model.get_package(pak_name), data=sp_data)
@staticmethod
def masked4D_arrays_to_stress_period_data(dtype, m4ds):
""" convert a dictionary of 4-dim masked arrays to
a stress_period_data style dict of recarray
Parameters
----------
dtype : numpy dtype
m4ds : dict {name:masked numpy 4-dim ndarray}
Returns
-------
dict {kper:recarray}
"""
assert isinstance(m4ds, dict)
for name, m4d in m4ds.items():
assert isinstance(m4d, np.ndarray)
assert name in dtype.names
assert m4d.ndim == 4
keys = list(m4ds.keys())
for i1, key1 in enumerate(keys):
a1 = np.isnan(m4ds[key1])
for i2, key2 in enumerate(keys[i1:]):
a2 = np.isnan(m4ds[key2])
if not np.array_equal(a1, a2):
raise Exception(
"Transient2d error: masking not equal"
+ " for {0} and {1}".format(key1, key2)
)
sp_data = {}
for kper in range(m4d.shape[0]):
vals = {}
for name, m4d in m4ds.items():
arr = m4d[kper, :, :, :]
isnan = np.argwhere(~np.isnan(arr))
v = []
for k, i, j in isnan:
v.append(arr[k, i, j])
vals[name] = v
kk = isnan[:, 0]
ii = isnan[:, 1]
jj = isnan[:, 2]
spd = np.recarray(shape=isnan.shape[0], dtype=dtype)
spd["i"] = ii
spd["k"] = kk
spd["j"] = jj
for n, v in vals.items():
spd[n] = v
sp_data[kper] = spd
return sp_data
| 34.349807 | 81 | 0.49012 | [
"CC0-1.0",
"BSD-3-Clause"
] | aleaf/flopy | flopy/utils/util_list.py | 44,483 | Python |
# -*- coding: utf-8 -*-
import requests
from webs.api.exceptions.customs import ServerError, InvalidAPIRequest, RecordNotFound, RecordAlreadyExists
class RequestMixin(object):
CODE_EXCEPTION_MSG = {
400: InvalidAPIRequest,
404: RecordNotFound,
409: RecordAlreadyExists,
422: InvalidAPIRequest,
500: ServerError,
}
def __init__(self):
self.session = requests.Session()
@property
def _headers(self):
return {
"Content-Type": "application/json",
}
def request(self, server, method, url, json=None, params=None, timeout=60):
try:
response = self.session.request(
method, url, json=json, params=params,
timeout=timeout, headers=self._headers
)
except requests.exceptions.ConnectTimeout:
raise self.CODE_EXCEPTION_MSG[500](f"{server}服务器连接超时!")
except requests.exceptions.ConnectionError:
raise self.CODE_EXCEPTION_MSG[500](f"{server}服务器连接错误!")
try:
response_data = response.json()
except Exception as e:
raise ServerError(f"{server}服务器参数解析失败!")
if not (200 <= response.status_code < 300):
exception = self.CODE_EXCEPTION_MSG[response.status_code] \
if response.status_code in self.CODE_EXCEPTION_MSG else self.CODE_EXCEPTION_MSG[400]
raise exception(f"{server} Response:{response_data.get('error').get('message')}")
return response_data
web_client = RequestMixin()
| 30.230769 | 107 | 0.632316 | [
"Apache-2.0"
] | c89758971/crawloop | services/engine/webs/core/requests/request.py | 1,624 | Python |
import os
import re
import codecs
def isValidLine(line):
if re.search('include \"', line) == None or line.find('.PSVita') != -1 or line.find('.PS4') != -1 or line.find('.Switch') != -1 or line.find('.XBoxOne') != -1:
return True
return False
class CreateHeader:
def __init__(self):
self.lines = []
def addLine(self,line):
self.lines.append(line)
def readLines(self,path):
f = codecs.open(path, 'r','utf-8_sig')
line = f.readline()
while line:
if isValidLine(line):
self.lines.append(line.strip(os.linesep))
line = f.readline()
f.close()
def output(self,path):
f = codecs.open(path, 'w','utf-8_sig')
for line in self.lines:
f.write(line + os.linesep)
f.close()
effekseerHeader = CreateHeader()
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Base.Pre.h')
effekseerHeader.readLines('Effekseer/Effekseer/Utils/Effekseer.CustomAllocator.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Vector2D.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Vector3D.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Color.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.RectF.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Matrix43.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Matrix44.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.File.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.DefaultFile.h')
effekseerHeader.readLines('Effekseer/Effekseer/Backend/GraphicsDevice.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Resource.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Effect.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Manager.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Setting.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Server.h')
effekseerHeader.readLines('Effekseer/Effekseer/Effekseer.Client.h')
effekseerHeader.addLine('')
effekseerHeader.addLine('#include "Effekseer.Modules.h"')
effekseerHeader.addLine('')
effekseerHeader.output('Effekseer/Effekseer.h')
effekseerSimdHeader = CreateHeader()
effekseerSimdHeader.addLine('#pragma once')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Base.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Float4_Gen.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Float4_NEON.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Float4_SSE.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Int4_Gen.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Int4_NEON.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Int4_SSE.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Bridge_Gen.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Bridge_NEON.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Bridge_SSE.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Vec2f.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Vec3f.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Vec4f.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Mat43f.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Mat44f.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Quaternionf.h')
effekseerSimdHeader.readLines('Effekseer/Effekseer/SIMD/Utils.h')
effekseerSimdHeader.output('Effekseer/Effekseer.SIMD.h')
effekseerModulesHeader = CreateHeader()
effekseerModulesHeader.addLine('#pragma once')
effekseerModulesHeader.addLine('')
effekseerModulesHeader.addLine('#include "Effekseer.h"')
effekseerModulesHeader.addLine('#include "Effekseer.SIMD.h"')
effekseerModulesHeader.addLine('')
effekseerModulesHeader.addLine('// A header to access internal data of effekseer')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Parameter/Effekseer.Parameters.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Renderer/Effekseer.SpriteRenderer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Renderer/Effekseer.RibbonRenderer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Renderer/Effekseer.RingRenderer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Renderer/Effekseer.ModelRenderer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Renderer/Effekseer.TrackRenderer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.EffectLoader.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.TextureLoader.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Model/Model.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Model/ModelLoader.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.MaterialLoader.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Model/Model.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.Curve.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.CurveLoader.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Sound/Effekseer.SoundPlayer.h')
effekseerModulesHeader.readLines('Effekseer/Effekseer/Effekseer.SoundLoader.h')
effekseerModulesHeader.output('Effekseer/Effekseer.Modules.h')
effekseerRendererDX9Header = CreateHeader()
effekseerRendererDX9Header.readLines('EffekseerRendererDX9/EffekseerRenderer/EffekseerRendererDX9.Base.Pre.h')
effekseerRendererDX9Header.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererDX9Header.readLines('EffekseerRendererDX9/EffekseerRenderer/EffekseerRendererDX9.Renderer.h')
effekseerRendererDX9Header.output('EffekseerRendererDX9/EffekseerRendererDX9.h')
effekseerRendererDX11Header = CreateHeader()
effekseerRendererDX11Header.readLines('EffekseerRendererDX11/EffekseerRenderer/EffekseerRendererDX11.Base.Pre.h')
effekseerRendererDX11Header.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererDX11Header.readLines('EffekseerRendererDX11/EffekseerRenderer/EffekseerRendererDX11.Renderer.h')
effekseerRendererDX11Header.output('EffekseerRendererDX11/EffekseerRendererDX11.h')
effekseerRendererDX12Header = CreateHeader()
effekseerRendererDX12Header.readLines('EffekseerRendererDX12/EffekseerRenderer/EffekseerRendererDX12.Base.Pre.h')
effekseerRendererDX12Header.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererDX12Header.readLines('EffekseerRendererDX12/EffekseerRenderer/EffekseerRendererDX12.Renderer.h')
effekseerRendererDX12Header.readLines('EffekseerRendererLLGI/Common.h')
effekseerRendererDX12Header.output('EffekseerRendererDX12/EffekseerRendererDX12.h')
effekseerRendererVulkanHeader = CreateHeader()
effekseerRendererVulkanHeader.readLines('EffekseerRendererVulkan/EffekseerRenderer/EffekseerRendererVulkan.Base.Pre.h')
effekseerRendererVulkanHeader.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererVulkanHeader.readLines('EffekseerRendererVulkan/EffekseerRenderer/EffekseerRendererVulkan.Renderer.h')
effekseerRendererVulkanHeader.readLines('EffekseerRendererLLGI/Common.h')
effekseerRendererVulkanHeader.output('EffekseerRendererVulkan/EffekseerRendererVulkan.h')
effekseerRendererGLHeader = CreateHeader()
effekseerRendererGLHeader.readLines('EffekseerRendererGL/EffekseerRenderer/EffekseerRendererGL.Base.Pre.h')
effekseerRendererGLHeader.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererGLHeader.readLines('EffekseerRendererGL/EffekseerRenderer/EffekseerRendererGL.Renderer.h')
effekseerRendererGLHeader.output('EffekseerRendererGL/EffekseerRendererGL.h')
effekseerRendererMetalHeader = CreateHeader()
effekseerRendererMetalHeader.readLines('EffekseerRendererMetal/EffekseerRenderer/EffekseerRendererMetal.Base.Pre.h')
effekseerRendererMetalHeader.readLines('EffekseerRendererCommon/EffekseerRenderer.Renderer.h')
effekseerRendererMetalHeader.readLines('EffekseerRendererMetal/EffekseerRenderer/EffekseerRendererMetal.Renderer.h')
effekseerRendererMetalHeader.readLines('EffekseerRendererLLGI/Common.h')
effekseerRendererMetalHeader.output('EffekseerRendererMetal/EffekseerRendererMetal.h')
| 57.757143 | 160 | 0.852461 | [
"Apache-2.0",
"BSD-3-Clause"
] | Shockblast/Effekseer | Dev/Cpp/CreateHeader.py | 8,086 | Python |
import wx.stc as stc
def highlight(editor, styles, faces):
editor.SetLexer(stc.STC_LEX_YAML)
editor.StyleSetSpec(stc.STC_YAML_DEFAULT, "fore:" +
styles["default"] + ",face:%(helv)s,size:%(size)d" % faces)
editor.StyleSetSpec(stc.STC_YAML_COMMENT, "fore:" +
styles["comment"] + ",face:%(helv)s,size:%(size)d" % faces)
editor.StyleSetSpec(stc.STC_YAML_ERROR, "fore:" +
styles["error"] + ",face:%(helv)s,size:%(size)d" % faces)
editor.StyleSetSpec(stc.STC_YAML_IDENTIFIER, "fore:" +
styles["identifier"] + ",face:%(helv)s,size:%(size)d" % faces)
editor.StyleSetSpec(stc.STC_YAML_NUMBER, "fore:" +
styles["number"] + ",face:%(helv)s,size:%(size)d" % faces)
| 35.043478 | 86 | 0.569479 | [
"MIT"
] | ShardulNalegave/pycode | yamlHighlighter.py | 806 | Python |
import pickle
import os
from tqdm import tqdm
with open('../data/bawe_splits.p', 'rb') as f:
splits = pickle.load(f)
if not os.path.isdir('../data/preprocess/bawe-group'):
os.mkdir('../data/preprocess/bawe-group')
for filename in tqdm(splits['train']):
id = filename[:4]
with open(f'../data/bawe/CORPUS_TXT/{filename}', 'r') as f:
if not os.path.isdir(f'../data/preprocess/bawe-group/{id}'):
os.mkdir(f'../data/preprocess/bawe-group/{id}')
text = f.read()
with open(f'../data/preprocess/bawe-group/{id}/{filename}', 'w') as wf:
wf.write(text)
| 29.142857 | 79 | 0.609477 | [
"Apache-2.0"
] | grchristensen/avpd | notebooks/develop/2021-02-18-gc-bawe-data-grouping.py | 612 | Python |
# -*- coding: utf-8 -*-
from flask import Blueprint, jsonify
from flask_service.swagger import spec
__all__ = ['main_app']
main_app = Blueprint('main_app', __name__)
@main_app.route('/api')
def swagger():
"""
Responds with the OpenAPI specification for this application.
"""
return jsonify(spec.to_dict())
@main_app.route('/health')
def health():
"""
Responds with the current's service health.
Could be used by the liveness probe of a Kubernetes cluster for instance.
"""
# put some logic here to decide if your app is doing well or not
# by default, we'll always return everything is okay!
return ""
@main_app.route('/status')
def status():
"""
Responds with the current's service status.
Could be used by the readiness probe of a Kubernetes cluster.
"""
# put some logic here to decide if your app is doing well or not
# by default, we'll always return everything is okay!
return "" | 24.2 | 77 | 0.677686 | [
"Apache-2.0"
] | mwprog/atomist-flask-microservice | flask_service/views.py | 968 | Python |
"""Principal Component Analysis Base Classes"""
# Author: Alexandre Gramfort <[email protected]>
# Olivier Grisel <[email protected]>
# Mathieu Blondel <[email protected]>
# Denis A. Engemann <[email protected]>
# Kyle Kastner <[email protected]>
#
# License: BSD 3 clause
import numpy as np
from scipy import linalg
from ..base import BaseEstimator, TransformerMixin, _ClassNamePrefixFeaturesOutMixin
from ..utils.validation import check_is_fitted
from abc import ABCMeta, abstractmethod
class _BasePCA(
_ClassNamePrefixFeaturesOutMixin, TransformerMixin, BaseEstimator, metaclass=ABCMeta
):
"""Base class for PCA methods.
Warning: This class should not be used directly.
Use derived classes instead.
"""
def get_covariance(self):
"""Compute data covariance with the generative model.
``cov = components_.T * S**2 * components_ + sigma2 * eye(n_features)``
where S**2 contains the explained variances, and sigma2 contains the
noise variances.
Returns
-------
cov : array of shape=(n_features, n_features)
Estimated covariance of data.
"""
components_ = self.components_
exp_var = self.explained_variance_
if self.whiten:
components_ = components_ * np.sqrt(exp_var[:, np.newaxis])
exp_var_diff = np.maximum(exp_var - self.noise_variance_, 0.0)
cov = np.dot(components_.T * exp_var_diff, components_)
cov.flat[:: len(cov) + 1] += self.noise_variance_ # modify diag inplace
return cov
def get_precision(self):
"""Compute data precision matrix with the generative model.
Equals the inverse of the covariance but computed with
the matrix inversion lemma for efficiency.
Returns
-------
precision : array, shape=(n_features, n_features)
Estimated precision of data.
"""
n_features = self.components_.shape[1]
# handle corner cases first
if self.n_components_ == 0:
return np.eye(n_features) / self.noise_variance_
if self.n_components_ == n_features:
return linalg.inv(self.get_covariance())
# Get precision using matrix inversion lemma
components_ = self.components_
exp_var = self.explained_variance_
if self.whiten:
components_ = components_ * np.sqrt(exp_var[:, np.newaxis])
exp_var_diff = np.maximum(exp_var - self.noise_variance_, 0.0)
precision = np.dot(components_, components_.T) / self.noise_variance_
precision.flat[:: len(precision) + 1] += 1.0 / exp_var_diff
precision = np.dot(components_.T, np.dot(linalg.inv(precision), components_))
precision /= -(self.noise_variance_ ** 2)
precision.flat[:: len(precision) + 1] += 1.0 / self.noise_variance_
return precision
@abstractmethod
def fit(self, X, y=None):
"""Placeholder for fit. Subclasses should implement this method!
Fit the model with X.
Parameters
----------
X : array-like of shape (n_samples, n_features)
Training data, where `n_samples` is the number of samples and
`n_features` is the number of features.
Returns
-------
self : object
Returns the instance itself.
"""
def transform(self, X):
"""Apply dimensionality reduction to X.
X is projected on the first principal components previously extracted
from a training set.
Parameters
----------
X : array-like of shape (n_samples, n_features)
New data, where `n_samples` is the number of samples
and `n_features` is the number of features.
Returns
-------
X_new : array-like of shape (n_samples, n_components)
Projection of X in the first principal components, where `n_samples`
is the number of samples and `n_components` is the number of the components.
"""
check_is_fitted(self)
X = self._validate_data(X, dtype=[np.float64, np.float32], reset=False)
if self.mean_ is not None:
X = X - self.mean_
X_transformed = np.dot(X, self.components_.T)
if self.whiten:
X_transformed /= np.sqrt(self.explained_variance_)
return X_transformed
def inverse_transform(self, X):
"""Transform data back to its original space.
In other words, return an input `X_original` whose transform would be X.
Parameters
----------
X : array-like of shape (n_samples, n_components)
New data, where `n_samples` is the number of samples
and `n_components` is the number of components.
Returns
-------
X_original array-like of shape (n_samples, n_features)
Original data, where `n_samples` is the number of samples
and `n_features` is the number of features.
Notes
-----
If whitening is enabled, inverse_transform will compute the
exact inverse operation, which includes reversing whitening.
"""
if self.whiten:
return (
np.dot(
X,
np.sqrt(self.explained_variance_[:, np.newaxis]) * self.components_,
)
+ self.mean_
)
else:
return np.dot(X, self.components_) + self.mean_
@property
def _n_features_out(self):
"""Number of transformed output features."""
return self.components_.shape[0]
| 34.853659 | 88 | 0.618964 | [
"BSD-3-Clause"
] | 40104/Scikit-Learn | sklearn/decomposition/_base.py | 5,716 | Python |
from setuptools import setup
from os import path
from sys import version_info
def open_file(fname):
return open(path.join(path.dirname(__file__), fname))
setup_requires = ['pbr']
setup(
license='MIT',
setup_requires=setup_requires,
pbr=True,
auto_version="PBR",
install_requires=open(path.join(path.dirname(__file__), 'requirements.txt')).readlines(),
)
| 21.222222 | 93 | 0.727749 | [
"MIT"
] | CuidaAdmin/django-alexa | setup.py | 382 | Python |
"""
enCount tasks and analyses.
enCount is a Python library for processing RNA-Seq data from ENCODE.
"""
# from ._version import __version__
from . import config # load from myconfig.py if it exists
from . import db
from . import queues
from . import encode
from . import externals
from . import gtfs
from . import fastqs
from . import experiments
from . import mappings
from . import integration | 19.190476 | 68 | 0.756824 | [
"MIT"
] | mstrazar/enCount | enCount/__init__.py | 403 | Python |
from asgiref.sync import sync_to_async
from channels.layers import get_channel_layer
from ....models import Participant
import humps
channel_layer = get_channel_layer()
def get_participant(room_channel_name, channel_name):
participant = Participant.objects.get(
channel_room__channel_name=room_channel_name,
channel_name=channel_name
)
return participant
def get_participant_id(participant):
return participant.id
async def broadcast_avatar_position(room_channel_name, channel_name, json_data):
"""
Sends the new avatar's position to the users of the room.
"""
type = json_data['type']
payload = json_data['payload']
position = payload["position"]
animate = payload["animate"]
# receive the participant that sent this message
participant = await sync_to_async(get_participant)(room_channel_name, channel_name)
participant_id = await sync_to_async(get_participant_id)(participant)
# if this was for an avatar, then set participant's position to the payload data
def set_participant_position():
participant.x = position["x"]
participant.y = position["y"]
participant.direction_x = position["directionX"]
participant.save()
await sync_to_async(set_participant_position)()
await channel_layer.group_send(
room_channel_name,
{
'type': type,
'payload': {
"participant_id": participant_id,
"position": position,
"animate": animate,
}
}
)
async def broadcast_avatar_state(room_channel_name, channel_name, json_data):
"""
Sends the new avatar's state to the users of the room.
"""
type = json_data['type']
payload = json_data['payload']
state = payload['value']
# receive the participant that sent this message
participant = await sync_to_async(get_participant)(room_channel_name, channel_name)
participant_id = await sync_to_async(get_participant_id)(participant)
await channel_layer.group_send(
room_channel_name,
{
'type': humps.decamelize(type),
'payload': {
"participant_id": participant_id,
"state": state
}
}
) | 30.573333 | 87 | 0.66812 | [
"MIT"
] | Shadowsych/html5-msoy | server/websockets/consumers/world/broadcasts/avatar.py | 2,293 | Python |
"""Plot graphs from human-readable file formats."""
| 26 | 51 | 0.730769 | [
"MIT"
] | Sean1708/uniplot | uniplot/__init__.py | 52 | Python |
# https://www.hackerrank.com/challenges/tree-height-of-a-binary-tree/problem
def height(root):
"""
DFS
v = Vertices
e = Edges
d = Depth
Time complexity: O(v + e)
Space complexity: O(d)
"""
if root:
return 1 + max(height(root.left), height(root.right))
else:
return -1
| 17.421053 | 76 | 0.567976 | [
"MIT"
] | danielfsousa/algorithms-solutions | HackerRank/Data Structures/Trees/height-of-a-binary-tree.py | 331 | Python |
from dbt.clients.system import load_file_contents
from dbt.contracts.files import (
FilePath, ParseFileType, SourceFile, FileHash, AnySourceFile, SchemaSourceFile
)
from dbt.parser.schemas import yaml_from_file
from dbt.parser.search import FilesystemSearcher
# This loads the files contents and creates the SourceFile object
def load_source_file(
path: FilePath, parse_file_type: ParseFileType,
project_name: str) -> AnySourceFile:
file_contents = load_file_contents(path.absolute_path, strip=False)
checksum = FileHash.from_contents(file_contents)
sf_cls = SchemaSourceFile if parse_file_type == ParseFileType.Schema else SourceFile
source_file = sf_cls(path=path, checksum=checksum,
parse_file_type=parse_file_type, project_name=project_name)
source_file.contents = file_contents.strip()
if parse_file_type == ParseFileType.Schema:
source_file.dfy = yaml_from_file(source_file)
return source_file
# Special processing for big seed files
def load_seed_source_file(match: FilePath, project_name) -> SourceFile:
if match.seed_too_large():
# We don't want to calculate a hash of this file. Use the path.
source_file = SourceFile.big_seed(match)
else:
file_contents = load_file_contents(match.absolute_path, strip=False)
checksum = FileHash.from_contents(file_contents)
source_file = SourceFile(path=match, checksum=checksum)
source_file.contents = ''
source_file.parse_file_type = ParseFileType.Seed
source_file.project_name = project_name
return source_file
# Use the FilesystemSearcher to get a bunch of FilePaths, then turn
# them into a bunch of FileSource objects
def get_source_files(project, paths, extension, parse_file_type):
# file path list
fp_list = list(FilesystemSearcher(
project, paths, extension
))
# file block list
fb_list = []
for fp in fp_list:
if parse_file_type == ParseFileType.Seed:
fb_list.append(load_seed_source_file(fp, project.project_name))
else:
fb_list.append(load_source_file(
fp, parse_file_type, project.project_name))
return fb_list
def read_files_for_parser(project, files, dirs, extension, parse_ft):
parser_files = []
source_files = get_source_files(
project, dirs, extension, parse_ft
)
for sf in source_files:
files[sf.file_id] = sf
parser_files.append(sf.file_id)
return parser_files
# This needs to read files for multiple projects, so the 'files'
# dictionary needs to be passed in. What determines the order of
# the various projects? Is the root project always last? Do the
# non-root projects need to be done separately in order?
def read_files(project, files, parser_files):
project_files = {}
project_files['MacroParser'] = read_files_for_parser(
project, files, project.macro_paths, '.sql', ParseFileType.Macro,
)
project_files['ModelParser'] = read_files_for_parser(
project, files, project.source_paths, '.sql', ParseFileType.Model,
)
project_files['SnapshotParser'] = read_files_for_parser(
project, files, project.snapshot_paths, '.sql', ParseFileType.Snapshot,
)
project_files['AnalysisParser'] = read_files_for_parser(
project, files, project.analysis_paths, '.sql', ParseFileType.Analysis,
)
project_files['DataTestParser'] = read_files_for_parser(
project, files, project.test_paths, '.sql', ParseFileType.Test,
)
project_files['SeedParser'] = read_files_for_parser(
project, files, project.data_paths, '.csv', ParseFileType.Seed,
)
project_files['DocumentationParser'] = read_files_for_parser(
project, files, project.docs_paths, '.md', ParseFileType.Documentation,
)
project_files['SchemaParser'] = read_files_for_parser(
project, files, project.all_source_paths, '.yml', ParseFileType.Schema,
)
# Also read .yaml files for schema files. Might be better to change
# 'read_files_for_parser' accept an array in the future.
yaml_files = read_files_for_parser(
project, files, project.all_source_paths, '.yaml', ParseFileType.Schema,
)
project_files['SchemaParser'].extend(yaml_files)
# Store the parser files for this particular project
parser_files[project.project_name] = project_files
| 37.457627 | 88 | 0.722398 | [
"Apache-2.0"
] | JLDLaughlin/dbt | core/dbt/parser/read_files.py | 4,420 | Python |
from typing import List
'''
1. subproblems: dp(amount) the minimum number of coins needed to make changes for amount of S using the given coin denomination
2. guessing: all the available denomination c_i
3. relate subproblems: dp(amount) = min(dp(amount - c_i) + 1) for all possible c_i
Time complexity: O(#subproblems * #coins)
'''
class Solution:
# top down solution
def coinChange(self, coins: List[int], amount: int) -> int:
# for amount less than 1, return 0
if amount < 1:
return 0
memo = {}
def helper(coins, amount):
# for subproblems that we have alreay solve and memorized
if amount in memo:
return memo[amount]
# base case, we reach out the bottom of the tree.
if amount == 0:
return 0
# go through all possible coin denomination(breaches in tree)
dp = float('inf')
for coin in coins:
if coin > amount:
continue
# relate subproblems
dp = min(helper(coins, amount - coin) + 1, dp)
memo[amount] = dp
return dp
helper(coins, amount)
return -1 if memo[amount] == float('inf') else memo[amount]
# bottom-up solution, DAG
def coinChange_2(self, coins: List[int], amount: int) -> int:
memo = [float('inf') for i in range(amount + 1)]
# dp[i] = min{dp[i - c_i] + 1} for all c_i
memo[0] = 0
for i in range(amount + 1):
# check all the states that are reachable by coins to state i
for coin in coins:
if i < coin:
continue
memo[i] = min(memo[i], memo[i - coin] + 1)
print(memo)
return -1 if memo[amount] == float('inf') else memo[amount]
x = Solution()
# rs = x.coinChange([1, 2, 5], 2)
print(x.coinChange_2([1,2,5], 11)) | 28.109589 | 127 | 0.520955 | [
"MIT"
] | sundaycat/Leetcode-Practice | solution/322. coin-change.py | 2,052 | Python |
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'Raster Vector Analysis',
'author': 'Jan Kumor',
'url': 'http://github.com/akumor/python-rastervectoranalysis',
'download_url': 'http://github.com/akumor/python-rastervectoranalysis',
'author_email': '[email protected]',
'version': '0.1',
'install_requires': [''],
'packages': ['rastervectoranalysis'],
'scripts': [],
'name': 'rastervectoranalysis'
}
setup(**config)
| 27.2 | 75 | 0.667279 | [
"Apache-2.0"
] | akumor/python-rastervectoranalysis | setup.py | 544 | Python |
from checkov.common.models.enums import CheckCategories
from checkov.terraform.checks.resource.base_resource_value_check import BaseResourceValueCheck
class PasswordPolicyLowercaseLetter(BaseResourceValueCheck):
def __init__(self):
name = "Ensure RAM password policy requires at least one lowercase letter"
id = "CKV_ALI_17"
supported_resources = ['alicloud_ram_account_password_policy']
categories = [CheckCategories.IAM]
super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)
def get_inspected_key(self):
return 'require_lowercase_characters'
check = PasswordPolicyLowercaseLetter()
| 38.333333 | 106 | 0.778261 | [
"Apache-2.0"
] | Eliran-Turgeman/checkov | checkov/terraform/checks/resource/alicloud/PasswordPolicyLowercaseLetter.py | 690 | Python |
import SimpleITK as sitk
import numpy as np
import torch
import math
import time
import sys
import cv2
from scipy.ndimage.interpolation import zoom
from torch.autograd import Variable
sys.path.append('../lung_nodule_detector')
from training.layers import nms
def load_itk_image(filename):
with open(filename) as f:
contents = f.readlines()
line = [k for k in contents if k.startswith('TransformMatrix')][0]
transformM = np.array(line.split(' = ')[1].split(' ')).astype('float')
transformM = np.round(transformM)
if np.any(transformM != np.array([1, 0, 0, 0, 1, 0, 0, 0, 1])):
isflip = True
else:
isflip = False
itkimage = sitk.ReadImage(filename)
numpyImage = sitk.GetArrayFromImage(itkimage)
numpyOrigin = np.array(list(reversed(itkimage.GetOrigin())))
numpySpacing = np.array(list(reversed(itkimage.GetSpacing())))
return numpyImage, numpyOrigin, numpySpacing, isflip
def lumTrans(img):
lungwin = np.array([-1200.,600.])
newimg = (img-lungwin[0])/(lungwin[1]-lungwin[0])
newimg[newimg<0]=0
newimg[newimg>1]=1
newimg = (newimg*255).astype('uint8')
return newimg
def resample(imgs, spacing, new_spacing, progressBar, order=2):
print (len(imgs.shape))
if len(imgs.shape)==3:
new_shape = np.round(imgs.shape * spacing / new_spacing)
true_spacing = spacing * imgs.shape / new_shape
resize_factor = new_shape / imgs.shape
imgs = zoom(imgs, resize_factor, mode = 'nearest',order=order)
progressBar.setValue(40)
return imgs, true_spacing
elif len(imgs.shape)==4:
n = imgs.shape[-1]
newimg = []
for i in range(n):
slice = imgs[:,:,:,i]
newslice,true_spacing = resample(slice,spacing,new_spacing)
newimg.append(newslice)
newimg=np.transpose(np.array(newimg),[1,2,3,0])
return newimg,true_spacing
else:
raise ValueError('wrong shape')
def resample_v1(imgs, spacing, new_spacing, order=2):
print (len(imgs.shape))
if len(imgs.shape)==3:
new_shape = np.round(imgs.shape * spacing / new_spacing)
true_spacing = spacing * imgs.shape / new_shape
resize_factor = new_shape / imgs.shape
imgs = zoom(imgs, resize_factor, mode = 'nearest',order=order)
return imgs, true_spacing
elif len(imgs.shape)==4:
n = imgs.shape[-1]
newimg = []
for i in range(n):
slice = imgs[:,:,:,i]
newslice,true_spacing = resample(slice,spacing,new_spacing)
newimg.append(newslice)
newimg=np.transpose(np.array(newimg),[1,2,3,0])
return newimg,true_spacing
else:
raise ValueError('wrong shape')
def split_data(data, stride, split_comber):
print (data.shape[1:])
nz, nh, nw = data.shape[1:]
pz = int(np.ceil(float(nz) / stride)) * stride
ph = int(np.ceil(float(nh) / stride)) * stride
pw = int(np.ceil(float(nw) / stride)) * stride
data = np.pad(data, [[0, 0], [0, pz - nz], [0, ph - nh], [0, pw - nw]], 'constant', constant_values=0)
xx, yy, zz = np.meshgrid(np.linspace(-0.5, 0.5, data.shape[1] / stride),
np.linspace(-0.5, 0.5, data.shape[2] / stride),
np.linspace(-0.5, 0.5, data.shape[3] / stride), indexing='ij')
coord = np.concatenate([xx[np.newaxis, ...], yy[np.newaxis, ...], zz[np.newaxis, :]], 0).astype('float32')
data, nzhw = split_comber.split(data)
coord2, nzhw2 = split_comber.split(coord,
side_len=split_comber.side_len / stride,
max_stride=split_comber.max_stride / stride,
margin=split_comber.margin / stride)
assert np.all(nzhw == nzhw2)
data = (data.astype(np.float32) - 128) / 128
return torch.from_numpy(data), torch.from_numpy(coord2), np.array(nzhw)
def convert_prob(pbb):
for label in pbb:
pos_ori = label[1:4]
radious_ori = label[4]
#pos_ori = pos_ori + extendbox[:, 0]
label[1:4] = pos_ori
label[4] = radious_ori
label[0] = sigmoid(label[0])
return pbb
def sigmoid(x):
return 1 / (1 + math.exp(-x))
def predict_nodule(net, data, coord, nzhw, lbb, n_per_run, split_comber, get_pbb, progressBar):
net.eval()
total_label = 0
total_candi = 0
splitlist = list(range(0, len(data) + 1, n_per_run))
if splitlist[-1] != len(data):
splitlist.append(len(data))
outputlist = []
for i in range(len(splitlist) - 1):
with torch.no_grad():
inputdata = Variable(data[splitlist[i]:splitlist[i + 1]]).cuda()
inputcoord = Variable(coord[splitlist[i]:splitlist[i + 1]]).cuda()
output = net(inputdata, inputcoord)
outputlist.append(output.data.cpu().numpy())
progressBar.setValue(10 + (80/len(splitlist) * (i+1)))
output = np.concatenate(outputlist, 0)
output = split_comber.combine(output, nzhw=nzhw)
# fps 1.215909091, sens 0.933333333, thres 0.371853054
thresh = 0.371853054
pbb, mask = get_pbb(output, thresh, ismask=True)
pbb = pbb[pbb[:, 0].argsort()[::-1]]
pbb_cand_list = []
# check overlap under 3mm
for cand in pbb:
is_overlap = False
for appended in pbb_cand_list:
minimum_dist = 3
dist = math.sqrt(
math.pow(appended[1] - cand[1], 2) + math.pow(appended[2] - cand[2], 2) + math.pow(
appended[3] - cand[3], 2))
if (dist < minimum_dist):
is_overlap = True
break;
if not is_overlap:
pbb_cand_list.append(cand)
pbb_cand_list = np.array(pbb_cand_list)
pbb_cand_list_nms = nms(pbb_cand_list, 0.3)
# print (name)
# print (lbb)
world_pbb = convert_prob(pbb_cand_list_nms)
# print (world_pbb)
print("label", len(lbb))
print("z_pos y_pos x_pos size")
for i in range(len(lbb)):
for j in range(len(lbb[i])):
print(round(lbb[i][j], 2), end='\t')
print()
print("candidate", len(world_pbb))
print("prob z_pos y_pos x_pos size")
for i in range(len(world_pbb)):
for j in range(len(world_pbb[i])):
print(round(world_pbb[i][j], 2), end='\t')
print()
total_label += len(lbb)
total_candi += len(world_pbb)
return lbb, world_pbb
def predict_nodule_v1(net, data, coord, nzhw, lbb, n_per_run, split_comber, get_pbb):
net.eval()
total_label = 0
total_candi = 0
splitlist = list(range(0, len(data) + 1, n_per_run))
if splitlist[-1] != len(data):
splitlist.append(len(data))
outputlist = []
for i in range(len(splitlist) - 1):
with torch.no_grad():
inputdata = Variable(data[splitlist[i]:splitlist[i + 1]]).cuda()
inputcoord = Variable(coord[splitlist[i]:splitlist[i + 1]]).cuda()
output = net(inputdata, inputcoord)
outputlist.append(output.data.cpu().numpy())
output = np.concatenate(outputlist, 0)
output = split_comber.combine(output, nzhw=nzhw)
# fps 1.215909091, sens 0.933333333, thres 0.371853054
thresh = 0.371853054
pbb, mask = get_pbb(output, thresh, ismask=True)
pbb = pbb[pbb[:, 0].argsort()[::-1]]
pbb_cand_list = []
# check overlap under 3mm
for cand in pbb:
is_overlap = False
for appended in pbb_cand_list:
minimum_dist = 3
dist = math.sqrt(
math.pow(appended[1] - cand[1], 2) + math.pow(appended[2] - cand[2], 2) + math.pow(
appended[3] - cand[3], 2))
if (dist < minimum_dist):
is_overlap = True
break;
if not is_overlap:
pbb_cand_list.append(cand)
pbb_cand_list = np.array(pbb_cand_list)
pbb_cand_list_nms = nms(pbb_cand_list, 0.3)
# print (name)
# print (lbb)
world_pbb = convert_prob(pbb_cand_list_nms)
# print (world_pbb)
print("label", len(lbb))
print("z_pos y_pos x_pos size")
for i in range(len(lbb)):
for j in range(len(lbb[i])):
print(round(lbb[i][j], 2), end='\t')
print()
print("candidate", len(world_pbb))
print("prob z_pos y_pos x_pos size")
for i in range(len(world_pbb)):
for j in range(len(world_pbb[i])):
print(round(world_pbb[i][j], 2), end='\t')
print()
total_label += len(lbb)
total_candi += len(world_pbb)
return lbb, world_pbb
def draw_nodule_rect(lbb, world_pbb, img_arr):
for i in range(len(lbb)):
label = lbb[i]
# label = np.ceil(label)
r = (label[3] / 2) * 1.3
top_left = (max(int(math.ceil(label[2] - r)), 0),
max(int(math.ceil(label[1] - r)), 0))
bottom_right = (min(int(math.ceil(label[2] + r)), np.shape(img_arr)[1]),
min(int(math.ceil(label[1] + r)), np.shape(img_arr)[2]))
z_range = [max(int(math.ceil(label[0] - r)), 0),
min(int(math.ceil(label[0] + r)), np.shape(img_arr)[0])]
for j in range(z_range[0], z_range[1]):
cv2.rectangle(img_arr[j], top_left, bottom_right, (0, 255, 0), 1)
for i in range(len(world_pbb)):
candidate = world_pbb[i]
r = (candidate[4] / 2) * 1.3
top_left = (max(int(math.ceil(candidate[3] - r)), 0),
max(int(math.ceil(candidate[2] - r)), 0))
text_top_left = (max(int(math.ceil(candidate[3] - r)) - 1, 0),
max(int(math.ceil(candidate[2] - r)) - 1, 0))
bottom_right = (min(int(math.ceil(candidate[3] + r)), np.shape(img_arr)[1]),
min(int(math.ceil(candidate[2] + r)), np.shape(img_arr)[2]))
z_range = [max(int(math.ceil(candidate[1] - r)), 0),
min(int(math.ceil(candidate[1] + r)), np.shape(img_arr)[0])]
font = cv2.FONT_HERSHEY_SIMPLEX
for j in range(z_range[0], z_range[1]):
cv2.rectangle(img_arr[j], top_left, bottom_right, (255, 0, 0), 1)
#cv2.putText(img_arr[j], "c" + str(i) + "_" +str(round(candidate[0], 2)), top_left, font, 0.4, (255, 0, 0), 1, cv2.LINE_AA)
cv2.putText(img_arr[j], "c" + str(i), text_top_left, font, 0.4, (255, 0, 0), 1, cv2.LINE_AA)
def crop_all(target, img_arr, crop_size = 48):
target = np.copy(target)
start = []
for i in range(3):
start.append(int(round(target[i])) - int(crop_size / 2))
pad = []
pad.append([0, 0])
for i in range(3):
leftpad = max(0, -start[i])
rightpad = max(0, start[i] + crop_size - img_arr.shape[i + 1])
pad.append([leftpad, rightpad])
crop = img_arr[:,
max(start[0], 0):min(start[0] + crop_size, img_arr.shape[1]),
max(start[1], 0):min(start[1] + crop_size, img_arr.shape[2]),
max(start[2], 0):min(start[2] + crop_size, img_arr.shape[3])]
crop = np.pad(crop, pad, 'constant', constant_values=0)
for i in range(3):
target[i] = target[i] - start[i]
return crop, target
def crop_nodule_arr_2ch(target, img_arr, crop_size = 48):
img_size = [crop_size, crop_size, crop_size]
crop_img, target = crop_all(target, img_arr, crop_size)
imgs = np.squeeze(crop_img, axis=0)
z = int(target[0])
y = int(target[1])
x = int(target[2])
print (z, y, x)
# z = 24
# y = 24
# x = 24
nodule_size = int(target[3])
margin = max(7, nodule_size * 0.4)
radius = int((nodule_size + margin) / 2)
s_z_pad = 0
e_z_pad = 0
s_y_pad = 0
e_y_pad = 0
s_x_pad = 0
e_x_pad = 0
s_z = max(0, z - radius)
if (s_z == 0):
s_z_pad = -(z - radius)
e_z = min(np.shape(imgs)[0], z + radius)
if (e_z == np.shape(imgs)[0]):
e_z_pad = (z + radius) - np.shape(imgs)[0]
s_y = max(0, y - radius)
if (s_y == 0):
s_y_pad = -(y - radius)
e_y = min(np.shape(imgs)[1], y + radius)
if (e_y == np.shape(imgs)[1]):
e_y_pad = (y + radius) - np.shape(imgs)[1]
s_x = max(0, x - radius)
if (s_x == 0):
s_x_pad = -(x - radius)
e_x = min(np.shape(imgs)[2], x + radius)
if (e_x == np.shape(imgs)[2]):
e_x_pad = (x + radius) - np.shape(imgs)[2]
# print (s_x, e_x, s_y, e_y, s_z, e_z)
# print (np.shape(img_arr[s_z:e_z, s_y:e_y, s_x:e_x]))
nodule_img = imgs[s_z:e_z, s_y:e_y, s_x:e_x]
nodule_img = np.pad(nodule_img, [[s_z_pad, e_z_pad], [s_y_pad, e_y_pad], [s_x_pad, e_x_pad]], 'constant',
constant_values=0)
imgpad_size = [img_size[0] - np.shape(nodule_img)[0],
img_size[1] - np.shape(nodule_img)[1],
img_size[2] - np.shape(nodule_img)[2]]
imgpad = []
imgpad_left = [int(imgpad_size[0] / 2),
int(imgpad_size[1] / 2),
int(imgpad_size[2] / 2)]
imgpad_right = [int(imgpad_size[0] / 2),
int(imgpad_size[1] / 2),
int(imgpad_size[2] / 2)]
for i in range(3):
if (imgpad_size[i] % 2 != 0):
rand = np.random.randint(2)
if rand == 0:
imgpad.append([imgpad_left[i], imgpad_right[i] + 1])
else:
imgpad.append([imgpad_left[i] + 1, imgpad_right[i]])
else:
imgpad.append([imgpad_left[i], imgpad_right[i]])
padding_crop = np.pad(nodule_img, imgpad, 'constant', constant_values=0)
padding_crop = np.expand_dims(padding_crop, axis=0)
crop = np.concatenate((padding_crop, crop_img))
crop = (crop.astype(np.float32) - 128) / 128
return torch.from_numpy(crop), crop
def predict_attribute(attribute_net, crop_img):
attribute_net.eval()
with torch.no_grad():
crop_img = Variable(crop_img.cuda(async=True))
output = attribute_net(crop_img)
return output
| 34.55774 | 135 | 0.572129 | [
"MIT"
] | JiazeWang/lung_nodule_integ_viewer | UI_util.py | 14,065 | Python |
# Copyright (c) 2021 Sen Wu. All Rights Reserved.
"""Helper function to set random seed for reproducibility of models."""
import logging
import random
from typing import Optional
import numpy as np
import torch
logger = logging.getLogger(__name__)
def set_random_seed(seed: Optional[int] = None) -> None:
"""Set random seed for random, numpy, and pytorch.
Args:
seed: The random seed, defaults to `None` which select it randomly.
"""
max_value = np.iinfo(np.uint32).max
min_value = np.iinfo(np.uint32).min
try:
seed = int(seed)
logger.info(f"Set random seed to {seed}.")
except (TypeError, ValueError):
seed = random.randint(min_value, max_value)
logger.info(f"No random seed specified, randomly set random seed to {seed}.")
if not (min_value <= seed <= max_value):
new_seed = random.randint(min_value, max_value)
logger.info(
f"Random seed {seed} is not valid, randomly set random seed to {new_seed}."
)
seed = new_seed
# Set random seed for random
random.seed(seed)
# Set random seed for all numpy operations
np.random.seed(seed=seed)
# Set random seed for PyTorch
torch.manual_seed(seed)
| 27.555556 | 87 | 0.664516 | [
"MIT"
] | KeAWang/emmental | src/emmental/utils/seed.py | 1,240 | Python |
# -*- coding: utf-8 -*-
"""
Copyright (c) 2012 University of Oxford
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, --INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
"""
from rdfdatabank.lib.auth_entry import list_silos, list_usernames, list_group_usernames, add_silo, add_group_users
def sync_members(g):
# NOTE: g._register_silos() IS AN EXPENSIVE OPERATION.
# THIS FUNCTION IS EXPENSIVE AND SHOULD BE CALLED ONLY IF REALLY NECESSARY
#g = ag.granary
g.state.revert()
g._register_silos()
granary_list = g.silos
granary_list_database = list_silos()
usernames = list_usernames()
for silo in granary_list:
if not silo in granary_list_database:
add_silo(silo)
kw = g.describe_silo(silo)
#Get existing owners, admins, managers and submitters from silo metadata
owners = []
admins = []
managers = []
submitters = []
if 'administrators' in kw and kw['administrators']:
admins = [x.strip() for x in kw['administrators'].split(",") if x]
if 'managers' in kw and kw['managers']:
managers = [x.strip() for x in kw['managers'].split(",") if x]
if 'submitters' in kw and kw['submitters']:
submitters = [x.strip() for x in kw['submitters'].split(",") if x]
# Check users in silo metadata are valid users
owners = [x for x in owners if x in usernames]
admins = [x for x in admins if x in usernames]
managers = [x for x in managers if x in usernames]
submitters = [x for x in submitters if x in usernames]
#Synchronize members in silo metadata with users in database
d_admins = []
d_managers = []
d_sunbmitters = []
if silo in granary_list_database:
d_admins, d_managers, d_submitters = list_group_usernames(silo)
admins.extend(d_admins)
managers.extend(d_managers)
submitters.extend(d_submitters)
# Ensure users are listed just once in silo metadata and owner is superset
owners.extend(admins)
owners.extend(managers)
owners.extend(submitters)
admins = list(set(admins))
managers = list(set(managers))
submitters = list(set(submitters))
owners = list(set(owners))
# Add users in silo metadata to the database
new_silo_users = []
for a in admins:
if not a in d_admins:
new_silo_users.append((a, 'administrator'))
for a in managers:
if not a in d_managers:
new_silo_users.append((a, 'manager'))
for a in new_submitters:
if not a in d_submitters:
new_silo_users.append((a, 'submitter'))
if new_silo_users:
add_group_users(silo, new_silo_users)
#Write members into silo
kw['owners'] = ','.join(owners)
kw['administrators'] = ','.join(admins)
kw['managers'] = ','.join(managers)
kw['submitters'] = ','.join(submitters)
g.describe_silo(silo, **kw)
g.sync()
return
| 39.803922 | 114 | 0.652956 | [
"MIT"
] | dataflow/RDFDatabank | rdfdatabank/lib/data_sync.py | 4,060 | Python |
import mysql.connector
import json
import os
import requests
def getAllFindings(host, database, user, password, table, where):
db = mysql.connector.connect(host=host, database=database, user=user, password=password)
cursor = db.cursor()
cursor.execute("SELECT distinct findingCode, specimenOrganCode FROM " + table + " " + where)
return cursor.fetchall()
def getDrugs(api, filename):
if filename is None:
drugs = getDrugsMapping(api)
else:
if os.path.isfile(filename):
with open(filename, 'r') as drug_file:
drugs = json.loads(drug_file.read())
else:
drugs = getDrugsMapping(api)
with open(filename, 'w') as drug_file:
drug_file.write(json.dumps(drugs))
return drugs
def getDrugsMapping(api):
result = {}
clinicalCompounds = getClinicalCompounds(api)
preclinicalCompounds = getPreclinicalCompounds(api)
# iterate over the clinical and preclinical compounds and match them om inchiKey
for clinicalCompound in clinicalCompounds:
for preclinicalCompound in preclinicalCompounds:
if (clinicalCompound['inchiKey'] is not None) and (clinicalCompound['inchiKey'] == preclinicalCompound['inchiKey']):
inchiKey = clinicalCompound['inchiKey']
if inchiKey not in result:
result[inchiKey] = {
'inchiKey': inchiKey,
'clinicalName': clinicalCompound['name'],
'preclinicalName': preclinicalCompound['name']
}
result[inchiKey][preclinicalCompound['source']] = preclinicalCompound['findingIds']
result[inchiKey][clinicalCompound['source']] = clinicalCompound['findingIds']
return result
def getClinicalCompounds(api):
ct_compounds = api.ClinicalTrials().getAllCompounds();
for ct_compound in ct_compounds:
ct_compound['source'] = 'ClinicalTrials'
ml_compounds = api.Medline().getAllCompounds();
for ml_compound in ml_compounds:
ml_compound['source'] = 'Medline'
fa_compounds = api.Faers().getAllCompounds();
for fa_compound in fa_compounds:
fa_compound['source'] = 'Faers'
dm_compounds = api.DailyMed().getAllCompounds();
for dm_compound in dm_compounds:
dm_compound['source'] = 'DailyMed'
return ct_compounds + ml_compounds + fa_compounds + dm_compounds
def getPreclinicalCompounds(api):
et_compounds = api.eToxSys().getAllCompounds()
for et_compound in et_compounds:
et_compound['source'] = 'eToxSys'
return et_compounds
def getFindingsByIds(api, service, findingIds):
result = []
record_count = 0
query = {
"filter": {
"criteria": [
[
{
"field": {
"dataClassKey": "FINDING",
"name": "id"
},
"primitiveType": "Integer",
"comparisonOperator": "IN",
"values": None
},
]
]
},
"selectedFields": [
{
"dataClassKey": "FINDING",
"names": [
"id",
"specimenOrgan", "specimenOrganCode", "specimenOrganVocabulary",
"findingIdentifier", "finding", "findingCode", "findingVocabulary", "findingType",
"severity", "observation", "frequency",
"dose", "doseUnit",
"timepoint", "timepointUnit",
"treatmentRelated",
"compoundId",
"studyId",
"createdDate", "modifiedDate", "sex"
]
}
],
"offset": 0,
"limit": 500
}
for offset in range(0, len(findingIds), 500):
query['filter']['criteria'][0][0]['values'] = [{'value': findingId} for findingId in findingIds[offset:offset+500]]
r = requests.post(service.endpoint + 'query', verify=False, headers={"Authorization": f"Bearer {api.get_token()}"}, json=query, timeout=None)
if r.status_code == 200:
response = json.loads(r.text)
for record in response['resultData']['data']:
record['FINDING']['source'] = response['origin']
result.append(record['FINDING'])
elif r.status_code == 401:
api.reconnect()
continue
return result
| 35.744186 | 149 | 0.561483 | [
"MIT"
] | erikvanmulligen/etransafe-heatmap | Concordance/condordance_utils.py | 4,611 | Python |
"""
Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.
Permission is hereby granted, free of charge, to any person obtaining a copy of this
software and associated documentation files (the "Software"), to deal in the Software
without restriction, including without limitation the rights to use, copy, modify,
merge, publish, distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A
PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
"""
from cfnlint.rules.outputs.Name import Name # pylint: disable=E0401
from .. import BaseRuleTestCase
class TestName(BaseRuleTestCase):
"""Test template outputs Names"""
def setUp(self):
"""Setup"""
super(TestName, self).setUp()
self.collection.register(Name())
self.success_templates = [
'fixtures/templates/good/outputs/name.yaml'
]
def test_file_positive(self):
"""Test Positive"""
self.helper_file_positive()
def test_file_negative(self):
"""Test failure"""
self.helper_file_negative('fixtures/templates/bad/outputs/name.yaml', 1)
| 42.131579 | 87 | 0.723923 | [
"MIT-0"
] | SanderKnape/cfn-python-lint | test/rules/outputs/test_name.py | 1,601 | Python |
# --------------------------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# --------------------------------------------------------------------------------------------
# pylint: disable=line-too-long
from ._client_factory import iotcentral_service_factory
def load_command_table(self, _):
from azure.cli.core.commands import CliCommandType
iotcentral_sdk = CliCommandType(
operations_tmpl='azure.mgmt.iotcentral.operations#IoTCentaralOperations.{}'
)
update_custom_util = CliCommandType(
operations_tmpl='azure.cli.command_modules.iotcentral.custom#{}')
with self.command_group('iotcentral app', iotcentral_sdk, client_factory=iotcentral_service_factory) as g:
g.custom_command('create', 'iotcentral_app_create')
g.custom_command('list', 'iotcentral_app_list')
g.custom_command('show', 'iotcentral_app_get')
g.generic_update_command('update', getter_name='iotcentral_app_get',
setter_name='iotcentral_app_update', command_type=update_custom_util)
g.custom_command('delete', 'iotcentral_app_delete')
| 46.142857 | 110 | 0.632353 | [
"MIT"
] | 6paklata/azure-cli | src/azure-cli/azure/cli/command_modules/iotcentral/commands.py | 1,292 | Python |
# Copyright 2015 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
"""RNN helpers for TensorFlow models."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from tensorflow.python.framework import constant_op
from tensorflow.python.framework import dtypes
from tensorflow.python.framework import ops
from tensorflow.python.framework import tensor_shape
from tensorflow.python.ops import array_ops
from tensorflow.python.ops import control_flow_ops
from tensorflow.python.ops import math_ops
from tensorflow.python.ops import rnn_cell_impl
from tensorflow.python.ops import tensor_array_ops
from tensorflow.python.ops import variable_scope as vs
from tensorflow.python.util import nest
# pylint: disable=protected-access
_state_size_with_prefix = rnn_cell_impl._state_size_with_prefix
# pylint: enable=protected-access
def _infer_state_dtype(explicit_dtype, state):
"""Infer the dtype of an RNN state.
Args:
explicit_dtype: explicitly declared dtype or None.
state: RNN's hidden state. Must be a Tensor or a nested iterable containing
Tensors.
Returns:
dtype: inferred dtype of hidden state.
Raises:
ValueError: if `state` has heterogeneous dtypes or is empty.
"""
if explicit_dtype is not None:
return explicit_dtype
elif nest.is_sequence(state):
inferred_dtypes = [element.dtype for element in nest.flatten(state)]
if not inferred_dtypes:
raise ValueError("Unable to infer dtype from empty state.")
all_same = all([x == inferred_dtypes[0] for x in inferred_dtypes])
if not all_same:
raise ValueError(
"State has tensors of different inferred_dtypes. Unable to infer a "
"single representative dtype.")
return inferred_dtypes[0]
else:
return state.dtype
def _on_device(fn, device):
"""Build the subgraph defined by lambda `fn` on `device` if it's not None."""
if device:
with ops.device(device):
return fn()
else:
return fn()
# pylint: disable=unused-argument
def _rnn_step(
time, sequence_length, min_sequence_length, max_sequence_length,
zero_output, state, call_cell, state_size, skip_conditionals=False):
"""Calculate one step of a dynamic RNN minibatch.
Returns an (output, state) pair conditioned on the sequence_lengths.
When skip_conditionals=False, the pseudocode is something like:
if t >= max_sequence_length:
return (zero_output, state)
if t < min_sequence_length:
return call_cell()
# Selectively output zeros or output, old state or new state depending
# on if we've finished calculating each row.
new_output, new_state = call_cell()
final_output = np.vstack([
zero_output if time >= sequence_lengths[r] else new_output_r
for r, new_output_r in enumerate(new_output)
])
final_state = np.vstack([
state[r] if time >= sequence_lengths[r] else new_state_r
for r, new_state_r in enumerate(new_state)
])
return (final_output, final_state)
Args:
time: Python int, the current time step
sequence_length: int32 `Tensor` vector of size [batch_size]
min_sequence_length: int32 `Tensor` scalar, min of sequence_length
max_sequence_length: int32 `Tensor` scalar, max of sequence_length
zero_output: `Tensor` vector of shape [output_size]
state: Either a single `Tensor` matrix of shape `[batch_size, state_size]`,
or a list/tuple of such tensors.
call_cell: lambda returning tuple of (new_output, new_state) where
new_output is a `Tensor` matrix of shape `[batch_size, output_size]`.
new_state is a `Tensor` matrix of shape `[batch_size, state_size]`.
state_size: The `cell.state_size` associated with the state.
skip_conditionals: Python bool, whether to skip using the conditional
calculations. This is useful for `dynamic_rnn`, where the input tensor
matches `max_sequence_length`, and using conditionals just slows
everything down.
Returns:
A tuple of (`final_output`, `final_state`) as given by the pseudocode above:
final_output is a `Tensor` matrix of shape [batch_size, output_size]
final_state is either a single `Tensor` matrix, or a tuple of such
matrices (matching length and shapes of input `state`).
Raises:
ValueError: If the cell returns a state tuple whose length does not match
that returned by `state_size`.
"""
# Convert state to a list for ease of use
flat_state = nest.flatten(state)
flat_zero_output = nest.flatten(zero_output)
def _copy_one_through(output, new_output):
copy_cond = (time >= sequence_length)
return _on_device(
lambda: array_ops.where(copy_cond, output, new_output),
device=new_output.op.device)
def _copy_some_through(flat_new_output, flat_new_state):
# Use broadcasting select to determine which values should get
# the previous state & zero output, and which values should get
# a calculated state & output.
flat_new_output = [
_copy_one_through(zero_output, new_output)
for zero_output, new_output in zip(flat_zero_output, flat_new_output)]
flat_new_state = [
_copy_one_through(state, new_state)
for state, new_state in zip(flat_state, flat_new_state)]
return flat_new_output + flat_new_state
def _maybe_copy_some_through():
"""Run RNN step. Pass through either no or some past state."""
new_output, new_state = call_cell()
nest.assert_same_structure(state, new_state)
flat_new_state = nest.flatten(new_state)
flat_new_output = nest.flatten(new_output)
return control_flow_ops.cond(
# if t < min_seq_len: calculate and return everything
time < min_sequence_length, lambda: flat_new_output + flat_new_state,
# else copy some of it through
lambda: _copy_some_through(flat_new_output, flat_new_state))
# TODO(ebrevdo): skipping these conditionals may cause a slowdown,
# but benefits from removing cond() and its gradient. We should
# profile with and without this switch here.
if skip_conditionals:
# Instead of using conditionals, perform the selective copy at all time
# steps. This is faster when max_seq_len is equal to the number of unrolls
# (which is typical for dynamic_rnn).
new_output, new_state = call_cell()
nest.assert_same_structure(state, new_state)
new_state = nest.flatten(new_state)
new_output = nest.flatten(new_output)
final_output_and_state = _copy_some_through(new_output, new_state)
else:
empty_update = lambda: flat_zero_output + flat_state
final_output_and_state = control_flow_ops.cond(
# if t >= max_seq_len: copy all state through, output zeros
time >= max_sequence_length, empty_update,
# otherwise calculation is required: copy some or all of it through
_maybe_copy_some_through)
if len(final_output_and_state) != len(flat_zero_output) + len(flat_state):
raise ValueError("Internal error: state and output were not concatenated "
"correctly.")
final_output = final_output_and_state[:len(flat_zero_output)]
final_state = final_output_and_state[len(flat_zero_output):]
for output, flat_output in zip(final_output, flat_zero_output):
output.set_shape(flat_output.get_shape())
for substate, flat_substate in zip(final_state, flat_state):
substate.set_shape(flat_substate.get_shape())
final_output = nest.pack_sequence_as(
structure=zero_output, flat_sequence=final_output)
final_state = nest.pack_sequence_as(
structure=state, flat_sequence=final_state)
return final_output, final_state
def _reverse_seq(input_seq, lengths):
"""Reverse a list of Tensors up to specified lengths.
Args:
input_seq: Sequence of seq_len tensors of dimension (batch_size, n_features)
or nested tuples of tensors.
lengths: A `Tensor` of dimension batch_size, containing lengths for each
sequence in the batch. If "None" is specified, simply reverses
the list.
Returns:
time-reversed sequence
"""
if lengths is None:
return list(reversed(input_seq))
flat_input_seq = tuple(nest.flatten(input_) for input_ in input_seq)
flat_results = [[] for _ in range(len(input_seq))]
for sequence in zip(*flat_input_seq):
input_shape = tensor_shape.unknown_shape(
ndims=sequence[0].get_shape().ndims)
for input_ in sequence:
input_shape.merge_with(input_.get_shape())
input_.set_shape(input_shape)
# Join into (time, batch_size, depth)
s_joined = array_ops.stack(sequence)
# TODO(schuster, ebrevdo): Remove cast when reverse_sequence takes int32
if lengths is not None:
lengths = math_ops.to_int64(lengths)
# Reverse along dimension 0
s_reversed = array_ops.reverse_sequence(s_joined, lengths, 0, 1)
# Split again into list
result = array_ops.unstack(s_reversed)
for r, flat_result in zip(result, flat_results):
r.set_shape(input_shape)
flat_result.append(r)
results = [nest.pack_sequence_as(structure=input_, flat_sequence=flat_result)
for input_, flat_result in zip(input_seq, flat_results)]
return results
def bidirectional_dynamic_rnn(cell_fw, cell_bw, inputs, sequence_length=None,
initial_state_fw=None, initial_state_bw=None,
dtype=None, parallel_iterations=None,
swap_memory=False, time_major=False, scope=None):
"""Creates a dynamic version of bidirectional recurrent neural network.
Similar to the unidirectional case above (rnn) but takes input and builds
independent forward and backward RNNs. The input_size of forward and
backward cell must match. The initial state for both directions is zero by
default (but can be set optionally) and no intermediate states are ever
returned -- the network is fully unrolled for the given (passed in)
length(s) of the sequence(s) or completely unrolled if length(s) is not
given.
Args:
cell_fw: An instance of RNNCell, to be used for forward direction.
cell_bw: An instance of RNNCell, to be used for backward direction.
inputs: The RNN inputs.
If time_major == False (default), this must be a tensor of shape:
`[batch_size, max_time, input_size]`.
If time_major == True, this must be a tensor of shape:
`[max_time, batch_size, input_size]`.
[batch_size, input_size].
sequence_length: An int32/int64 vector, size `[batch_size]`,
containing the actual lengths for each of the sequences.
initial_state_fw: (optional) An initial state for the forward RNN.
This must be a tensor of appropriate type and shape
`[batch_size, cell_fw.state_size]`.
If `cell_fw.state_size` is a tuple, this should be a tuple of
tensors having shapes `[batch_size, s] for s in cell_fw.state_size`.
initial_state_bw: (optional) Same as for `initial_state_fw`, but using
the corresponding properties of `cell_bw`.
dtype: (optional) The data type for the initial states and expected output.
Required if initial_states are not provided or RNN states have a
heterogeneous dtype.
parallel_iterations: (Default: 32). The number of iterations to run in
parallel. Those operations which do not have any temporal dependency
and can be run in parallel, will be. This parameter trades off
time for space. Values >> 1 use more memory but take less time,
while smaller values use less memory but computations take longer.
swap_memory: Transparently swap the tensors produced in forward inference
but needed for back prop from GPU to CPU. This allows training RNNs
which would typically not fit on a single GPU, with very minimal (or no)
performance penalty.
time_major: The shape format of the `inputs` and `outputs` Tensors.
If true, these `Tensors` must be shaped `[max_time, batch_size, depth]`.
If false, these `Tensors` must be shaped `[batch_size, max_time, depth]`.
Using `time_major = True` is a bit more efficient because it avoids
transposes at the beginning and end of the RNN calculation. However,
most TensorFlow data is batch-major, so by default this function
accepts input and emits output in batch-major form.
dtype: (optional) The data type for the initial state. Required if
either of the initial states are not provided.
scope: VariableScope for the created subgraph; defaults to
"bidirectional_rnn"
Returns:
A tuple (outputs, output_states) where:
outputs: A tuple (output_fw, output_bw) containing the forward and
the backward rnn output `Tensor`.
If time_major == False (default),
output_fw will be a `Tensor` shaped:
`[batch_size, max_time, cell_fw.output_size]`
and output_bw will be a `Tensor` shaped:
`[batch_size, max_time, cell_bw.output_size]`.
If time_major == True,
output_fw will be a `Tensor` shaped:
`[max_time, batch_size, cell_fw.output_size]`
and output_bw will be a `Tensor` shaped:
`[max_time, batch_size, cell_bw.output_size]`.
It returns a tuple instead of a single concatenated `Tensor`, unlike
in the `bidirectional_rnn`. If the concatenated one is preferred,
the forward and backward outputs can be concatenated as
`tf.concat(outputs, 2)`.
output_states: A tuple (output_state_fw, output_state_bw) containing
the forward and the backward final states of bidirectional rnn.
Raises:
TypeError: If `cell_fw` or `cell_bw` is not an instance of `RNNCell`.
"""
# pylint: disable=protected-access
if not isinstance(cell_fw, rnn_cell_impl._RNNCell):
raise TypeError("cell_fw must be an instance of RNNCell")
if not isinstance(cell_bw, rnn_cell_impl._RNNCell):
raise TypeError("cell_bw must be an instance of RNNCell")
# pylint: enable=protected-access
with vs.variable_scope(scope or "bidirectional_rnn"):
# Forward direction
with vs.variable_scope("fw") as fw_scope:
output_fw, output_state_fw = dynamic_rnn(
cell=cell_fw, inputs=inputs, sequence_length=sequence_length,
initial_state=initial_state_fw, dtype=dtype,
parallel_iterations=parallel_iterations, swap_memory=swap_memory,
time_major=time_major, scope=fw_scope)
# Backward direction
if not time_major:
time_dim = 1
batch_dim = 0
else:
time_dim = 0
batch_dim = 1
with vs.variable_scope("bw") as bw_scope:
inputs_reverse = array_ops.reverse_sequence(
input=inputs, seq_lengths=sequence_length,
seq_dim=time_dim, batch_dim=batch_dim)
tmp, output_state_bw = dynamic_rnn(
cell=cell_bw, inputs=inputs_reverse, sequence_length=sequence_length,
initial_state=initial_state_bw, dtype=dtype,
parallel_iterations=parallel_iterations, swap_memory=swap_memory,
time_major=time_major, scope=bw_scope)
output_bw = array_ops.reverse_sequence(
input=tmp, seq_lengths=sequence_length,
seq_dim=time_dim, batch_dim=batch_dim)
outputs = (output_fw, output_bw)
output_states = (output_state_fw, output_state_bw)
return (outputs, output_states)
def dynamic_rnn(cell, inputs, sequence_length=None, initial_state=None,
dtype=None, parallel_iterations=None, swap_memory=False,
time_major=False, scope=None):
"""Creates a recurrent neural network specified by RNNCell `cell`.
This function is functionally identical to the function `rnn` above, but
performs fully dynamic unrolling of `inputs`.
Unlike `rnn`, the input `inputs` is not a Python list of `Tensors`, one for
each frame. Instead, `inputs` may be a single `Tensor` where
the maximum time is either the first or second dimension (see the parameter
`time_major`). Alternatively, it may be a (possibly nested) tuple of
Tensors, each of them having matching batch and time dimensions.
The corresponding output is either a single `Tensor` having the same number
of time steps and batch size, or a (possibly nested) tuple of such tensors,
matching the nested structure of `cell.output_size`.
The parameter `sequence_length` is optional and is used to copy-through state
and zero-out outputs when past a batch element's sequence length. So it's more
for correctness than performance, unlike in rnn().
Args:
cell: An instance of RNNCell.
inputs: The RNN inputs.
If `time_major == False` (default), this must be a `Tensor` of shape:
`[batch_size, max_time, ...]`, or a nested tuple of such
elements.
If `time_major == True`, this must be a `Tensor` of shape:
`[max_time, batch_size, ...]`, or a nested tuple of such
elements.
This may also be a (possibly nested) tuple of Tensors satisfying
this property. The first two dimensions must match across all the inputs,
but otherwise the ranks and other shape components may differ.
In this case, input to `cell` at each time-step will replicate the
structure of these tuples, except for the time dimension (from which the
time is taken).
The input to `cell` at each time step will be a `Tensor` or (possibly
nested) tuple of Tensors each with dimensions `[batch_size, ...]`.
sequence_length: (optional) An int32/int64 vector sized `[batch_size]`.
initial_state: (optional) An initial state for the RNN.
If `cell.state_size` is an integer, this must be
a `Tensor` of appropriate type and shape `[batch_size, cell.state_size]`.
If `cell.state_size` is a tuple, this should be a tuple of
tensors having shapes `[batch_size, s] for s in cell.state_size`.
dtype: (optional) The data type for the initial state and expected output.
Required if initial_state is not provided or RNN state has a heterogeneous
dtype.
parallel_iterations: (Default: 32). The number of iterations to run in
parallel. Those operations which do not have any temporal dependency
and can be run in parallel, will be. This parameter trades off
time for space. Values >> 1 use more memory but take less time,
while smaller values use less memory but computations take longer.
swap_memory: Transparently swap the tensors produced in forward inference
but needed for back prop from GPU to CPU. This allows training RNNs
which would typically not fit on a single GPU, with very minimal (or no)
performance penalty.
time_major: The shape format of the `inputs` and `outputs` Tensors.
If true, these `Tensors` must be shaped `[max_time, batch_size, depth]`.
If false, these `Tensors` must be shaped `[batch_size, max_time, depth]`.
Using `time_major = True` is a bit more efficient because it avoids
transposes at the beginning and end of the RNN calculation. However,
most TensorFlow data is batch-major, so by default this function
accepts input and emits output in batch-major form.
scope: VariableScope for the created subgraph; defaults to "rnn".
Returns:
A pair (outputs, state) where:
outputs: The RNN output `Tensor`.
If time_major == False (default), this will be a `Tensor` shaped:
`[batch_size, max_time, cell.output_size]`.
If time_major == True, this will be a `Tensor` shaped:
`[max_time, batch_size, cell.output_size]`.
Note, if `cell.output_size` is a (possibly nested) tuple of integers
or `TensorShape` objects, then `outputs` will be a tuple having the
same structure as `cell.output_size`, containing Tensors having shapes
corresponding to the shape data in `cell.output_size`.
state: The final state. If `cell.state_size` is an int, this
will be shaped `[batch_size, cell.state_size]`. If it is a
`TensorShape`, this will be shaped `[batch_size] + cell.state_size`.
If it is a (possibly nested) tuple of ints or `TensorShape`, this will
be a tuple having the corresponding shapes.
Raises:
TypeError: If `cell` is not an instance of RNNCell.
ValueError: If inputs is None or an empty list.
"""
# pylint: disable=protected-access
if not isinstance(cell, rnn_cell_impl._RNNCell):
raise TypeError("cell must be an instance of RNNCell")
# pylint: enable=protected-access
# By default, time_major==False and inputs are batch-major: shaped
# [batch, time, depth]
# For internal calculations, we transpose to [time, batch, depth]
flat_input = nest.flatten(inputs)
if not time_major:
# (B,T,D) => (T,B,D)
flat_input = tuple(array_ops.transpose(input_, [1, 0, 2])
for input_ in flat_input)
parallel_iterations = parallel_iterations or 32
if sequence_length is not None:
sequence_length = math_ops.to_int32(sequence_length)
if sequence_length.get_shape().ndims not in (None, 1):
raise ValueError(
"sequence_length must be a vector of length batch_size, "
"but saw shape: %s" % sequence_length.get_shape())
sequence_length = array_ops.identity( # Just to find it in the graph.
sequence_length, name="sequence_length")
# Create a new scope in which the caching device is either
# determined by the parent scope, or is set to place the cached
# Variable using the same placement as for the rest of the RNN.
with vs.variable_scope(scope or "rnn") as varscope:
if varscope.caching_device is None:
varscope.set_caching_device(lambda op: op.device)
input_shape = tuple(array_ops.shape(input_) for input_ in flat_input)
batch_size = input_shape[0][1]
for input_ in input_shape:
if input_[1].get_shape() != batch_size.get_shape():
raise ValueError("All inputs should have the same batch size")
if initial_state is not None:
state = initial_state
else:
if not dtype:
raise ValueError("If no initial_state is provided, dtype must be.")
state = cell.zero_state(batch_size, dtype)
def _assert_has_shape(x, shape):
x_shape = array_ops.shape(x)
packed_shape = array_ops.stack(shape)
return control_flow_ops.Assert(
math_ops.reduce_all(math_ops.equal(x_shape, packed_shape)),
["Expected shape for Tensor %s is " % x.name,
packed_shape, " but saw shape: ", x_shape])
if sequence_length is not None:
# Perform some shape validation
with ops.control_dependencies(
[_assert_has_shape(sequence_length, [batch_size])]):
sequence_length = array_ops.identity(
sequence_length, name="CheckSeqLen")
inputs = nest.pack_sequence_as(structure=inputs, flat_sequence=flat_input)
(outputs, final_state) = _dynamic_rnn_loop(
cell,
inputs,
state,
parallel_iterations=parallel_iterations,
swap_memory=swap_memory,
sequence_length=sequence_length,
dtype=dtype)
# Outputs of _dynamic_rnn_loop are always shaped [time, batch, depth].
# If we are performing batch-major calculations, transpose output back
# to shape [batch, time, depth]
if not time_major:
# (T,B,D) => (B,T,D)
flat_output = nest.flatten(outputs)
flat_output = [array_ops.transpose(output, [1, 0, 2])
for output in flat_output]
outputs = nest.pack_sequence_as(
structure=outputs, flat_sequence=flat_output)
return (outputs, final_state)
def _dynamic_rnn_loop(cell,
inputs,
initial_state,
parallel_iterations,
swap_memory,
sequence_length=None,
dtype=None):
"""Internal implementation of Dynamic RNN.
Args:
cell: An instance of RNNCell.
inputs: A `Tensor` of shape [time, batch_size, input_size], or a nested
tuple of such elements.
initial_state: A `Tensor` of shape `[batch_size, state_size]`, or if
`cell.state_size` is a tuple, then this should be a tuple of
tensors having shapes `[batch_size, s] for s in cell.state_size`.
parallel_iterations: Positive Python int.
swap_memory: A Python boolean
sequence_length: (optional) An `int32` `Tensor` of shape [batch_size].
dtype: (optional) Expected dtype of output. If not specified, inferred from
initial_state.
Returns:
Tuple `(final_outputs, final_state)`.
final_outputs:
A `Tensor` of shape `[time, batch_size, cell.output_size]`. If
`cell.output_size` is a (possibly nested) tuple of ints or `TensorShape`
objects, then this returns a (possibly nsted) tuple of Tensors matching
the corresponding shapes.
final_state:
A `Tensor`, or possibly nested tuple of Tensors, matching in length
and shapes to `initial_state`.
Raises:
ValueError: If the input depth cannot be inferred via shape inference
from the inputs.
"""
state = initial_state
assert isinstance(parallel_iterations, int), "parallel_iterations must be int"
state_size = cell.state_size
flat_input = nest.flatten(inputs)
flat_output_size = nest.flatten(cell.output_size)
# Construct an initial output
input_shape = array_ops.shape(flat_input[0])
time_steps = input_shape[0]
batch_size = input_shape[1]
inputs_got_shape = tuple(input_.get_shape().with_rank_at_least(3)
for input_ in flat_input)
const_time_steps, const_batch_size = inputs_got_shape[0].as_list()[:2]
for shape in inputs_got_shape:
if not shape[2:].is_fully_defined():
raise ValueError(
"Input size (depth of inputs) must be accessible via shape inference,"
" but saw value None.")
got_time_steps = shape[0].value
got_batch_size = shape[1].value
if const_time_steps != got_time_steps:
raise ValueError(
"Time steps is not the same for all the elements in the input in a "
"batch.")
if const_batch_size != got_batch_size:
raise ValueError(
"Batch_size is not the same for all the elements in the input.")
# Prepare dynamic conditional copying of state & output
def _create_zero_arrays(size):
size = _state_size_with_prefix(size, prefix=[batch_size])
return array_ops.zeros(
array_ops.stack(size), _infer_state_dtype(dtype, state))
flat_zero_output = tuple(_create_zero_arrays(output)
for output in flat_output_size)
zero_output = nest.pack_sequence_as(structure=cell.output_size,
flat_sequence=flat_zero_output)
if sequence_length is not None:
min_sequence_length = math_ops.reduce_min(sequence_length)
max_sequence_length = math_ops.reduce_max(sequence_length)
time = array_ops.constant(0, dtype=dtypes.int32, name="time")
with ops.name_scope("dynamic_rnn") as scope:
base_name = scope
def _create_ta(name, dtype):
return tensor_array_ops.TensorArray(dtype=dtype,
size=time_steps,
tensor_array_name=base_name + name)
output_ta = tuple(_create_ta("output_%d" % i,
_infer_state_dtype(dtype, state))
for i in range(len(flat_output_size)))
input_ta = tuple(_create_ta("input_%d" % i, flat_input[0].dtype)
for i in range(len(flat_input)))
input_ta = tuple(ta.unstack(input_)
for ta, input_ in zip(input_ta, flat_input))
def _time_step(time, output_ta_t, state):
"""Take a time step of the dynamic RNN.
Args:
time: int32 scalar Tensor.
output_ta_t: List of `TensorArray`s that represent the output.
state: nested tuple of vector tensors that represent the state.
Returns:
The tuple (time + 1, output_ta_t with updated flow, new_state).
"""
input_t = tuple(ta.read(time) for ta in input_ta)
# Restore some shape information
for input_, shape in zip(input_t, inputs_got_shape):
input_.set_shape(shape[1:])
input_t = nest.pack_sequence_as(structure=inputs, flat_sequence=input_t)
call_cell = lambda: cell(input_t, state)
if sequence_length is not None:
(output, new_state) = _rnn_step(
time=time,
sequence_length=sequence_length,
min_sequence_length=min_sequence_length,
max_sequence_length=max_sequence_length,
zero_output=zero_output,
state=state,
call_cell=call_cell,
state_size=state_size,
skip_conditionals=True)
else:
(output, new_state) = call_cell()
# Pack state if using state tuples
output = nest.flatten(output)
output_ta_t = tuple(
ta.write(time, out) for ta, out in zip(output_ta_t, output))
return (time + 1, output_ta_t, new_state)
_, output_final_ta, final_state = control_flow_ops.while_loop(
cond=lambda time, *_: time < time_steps,
body=_time_step,
loop_vars=(time, output_ta, state),
parallel_iterations=parallel_iterations,
swap_memory=swap_memory)
# Unpack final output if not using output tuples.
final_outputs = tuple(ta.stack() for ta in output_final_ta)
# Restore some shape information
for output, output_size in zip(final_outputs, flat_output_size):
shape = _state_size_with_prefix(
output_size, prefix=[const_time_steps, const_batch_size])
output.set_shape(shape)
final_outputs = nest.pack_sequence_as(
structure=cell.output_size, flat_sequence=final_outputs)
return (final_outputs, final_state)
def raw_rnn(cell, loop_fn,
parallel_iterations=None, swap_memory=False, scope=None):
"""Creates an `RNN` specified by RNNCell `cell` and loop function `loop_fn`.
**NOTE: This method is still in testing, and the API may change.**
This function is a more primitive version of `dynamic_rnn` that provides
more direct access to the inputs each iteration. It also provides more
control over when to start and finish reading the sequence, and
what to emit for the output.
For example, it can be used to implement the dynamic decoder of a seq2seq
model.
Instead of working with `Tensor` objects, most operations work with
`TensorArray` objects directly.
The operation of `raw_rnn`, in pseudo-code, is basically the following:
```python
time = tf.constant(0, dtype=tf.int32)
(finished, next_input, initial_state, _, loop_state) = loop_fn(
time=time, cell_output=None, cell_state=None, loop_state=None)
emit_ta = TensorArray(dynamic_size=True, dtype=initial_state.dtype)
state = initial_state
while not all(finished):
(output, cell_state) = cell(next_input, state)
(next_finished, next_input, next_state, emit, loop_state) = loop_fn(
time=time + 1, cell_output=output, cell_state=cell_state,
loop_state=loop_state)
# Emit zeros and copy forward state for minibatch entries that are finished.
state = tf.where(finished, state, next_state)
emit = tf.where(finished, tf.zeros_like(emit), emit)
emit_ta = emit_ta.write(time, emit)
# If any new minibatch entries are marked as finished, mark these.
finished = tf.logical_or(finished, next_finished)
time += 1
return (emit_ta, state, loop_state)
```
with the additional properties that output and state may be (possibly nested)
tuples, as determined by `cell.output_size` and `cell.state_size`, and
as a result the final `state` and `emit_ta` may themselves be tuples.
A simple implementation of `dynamic_rnn` via `raw_rnn` looks like this:
```python
inputs = tf.placeholder(shape=(max_time, batch_size, input_depth),
dtype=tf.float32)
sequence_length = tf.placeholder(shape=(batch_size,), dtype=tf.int32)
inputs_ta = tf.TensorArray(dtype=tf.float32, size=max_time)
inputs_ta = inputs_ta.unstack(inputs)
cell = tf.contrib.rnn.LSTMCell(num_units)
def loop_fn(time, cell_output, cell_state, loop_state):
emit_output = cell_output # == None for time == 0
if cell_output is None: # time == 0
next_cell_state = cell.zero_state(batch_size, tf.float32)
else:
next_cell_state = cell_state
elements_finished = (time >= sequence_length)
finished = tf.reduce_all(elements_finished)
next_input = tf.cond(
finished,
lambda: tf.zeros([batch_size, input_depth], dtype=tf.float32),
lambda: inputs_ta.read(time))
next_loop_state = None
return (elements_finished, next_input, next_cell_state,
emit_output, next_loop_state)
outputs_ta, final_state, _ = raw_rnn(cell, loop_fn)
outputs = outputs_ta.stack()
```
Args:
cell: An instance of RNNCell.
loop_fn: A callable that takes inputs
`(time, cell_output, cell_state, loop_state)`
and returns the tuple
`(finished, next_input, next_cell_state, emit_output, next_loop_state)`.
Here `time` is an int32 scalar `Tensor`, `cell_output` is a
`Tensor` or (possibly nested) tuple of tensors as determined by
`cell.output_size`, and `cell_state` is a `Tensor`
or (possibly nested) tuple of tensors, as determined by the `loop_fn`
on its first call (and should match `cell.state_size`).
The outputs are: `finished`, a boolean `Tensor` of
shape `[batch_size]`, `next_input`: the next input to feed to `cell`,
`next_cell_state`: the next state to feed to `cell`,
and `emit_output`: the output to store for this iteration.
Note that `emit_output` should be a `Tensor` or (possibly nested)
tuple of tensors with shapes and structure matching `cell.output_size`
and `cell_output` above. The parameter `cell_state` and output
`next_cell_state` may be either a single or (possibly nested) tuple
of tensors. The parameter `loop_state` and
output `next_loop_state` may be either a single or (possibly nested) tuple
of `Tensor` and `TensorArray` objects. This last parameter
may be ignored by `loop_fn` and the return value may be `None`. If it
is not `None`, then the `loop_state` will be propagated through the RNN
loop, for use purely by `loop_fn` to keep track of its own state.
The `next_loop_state` parameter returned may be `None`.
The first call to `loop_fn` will be `time = 0`, `cell_output = None`,
`cell_state = None`, and `loop_state = None`. For this call:
The `next_cell_state` value should be the value with which to initialize
the cell's state. It may be a final state from a previous RNN or it
may be the output of `cell.zero_state()`. It should be a
(possibly nested) tuple structure of tensors.
If `cell.state_size` is an integer, this must be
a `Tensor` of appropriate type and shape `[batch_size, cell.state_size]`.
If `cell.state_size` is a `TensorShape`, this must be a `Tensor` of
appropriate type and shape `[batch_size] + cell.state_size`.
If `cell.state_size` is a (possibly nested) tuple of ints or
`TensorShape`, this will be a tuple having the corresponding shapes.
The `emit_output` value may be either `None` or a (possibly nested)
tuple structure of tensors, e.g.,
`(tf.zeros(shape_0, dtype=dtype_0), tf.zeros(shape_1, dtype=dtype_1))`.
If this first `emit_output` return value is `None`,
then the `emit_ta` result of `raw_rnn` will have the same structure and
dtypes as `cell.output_size`. Otherwise `emit_ta` will have the same
structure, shapes (prepended with a `batch_size` dimension), and dtypes
as `emit_output`. The actual values returned for `emit_output` at this
initializing call are ignored. Note, this emit structure must be
consistent across all time steps.
parallel_iterations: (Default: 32). The number of iterations to run in
parallel. Those operations which do not have any temporal dependency
and can be run in parallel, will be. This parameter trades off
time for space. Values >> 1 use more memory but take less time,
while smaller values use less memory but computations take longer.
swap_memory: Transparently swap the tensors produced in forward inference
but needed for back prop from GPU to CPU. This allows training RNNs
which would typically not fit on a single GPU, with very minimal (or no)
performance penalty.
scope: VariableScope for the created subgraph; defaults to "rnn".
Returns:
A tuple `(emit_ta, final_state, final_loop_state)` where:
`emit_ta`: The RNN output `TensorArray`.
If `loop_fn` returns a (possibly nested) set of Tensors for
`emit_output` during initialization, (inputs `time = 0`,
`cell_output = None`, and `loop_state = None`), then `emit_ta` will
have the same structure, dtypes, and shapes as `emit_output` instead.
If `loop_fn` returns `emit_output = None` during this call,
the structure of `cell.output_size` is used:
If `cell.output_size` is a (possibly nested) tuple of integers
or `TensorShape` objects, then `emit_ta` will be a tuple having the
same structure as `cell.output_size`, containing TensorArrays whose
elements' shapes correspond to the shape data in `cell.output_size`.
`final_state`: The final cell state. If `cell.state_size` is an int, this
will be shaped `[batch_size, cell.state_size]`. If it is a
`TensorShape`, this will be shaped `[batch_size] + cell.state_size`.
If it is a (possibly nested) tuple of ints or `TensorShape`, this will
be a tuple having the corresponding shapes.
`final_loop_state`: The final loop state as returned by `loop_fn`.
Raises:
TypeError: If `cell` is not an instance of RNNCell, or `loop_fn` is not
a `callable`.
"""
# pylint: disable=protected-access
if not isinstance(cell, rnn_cell_impl._RNNCell):
raise TypeError("cell must be an instance of RNNCell")
# pylint: enable=protected-access
if not callable(loop_fn):
raise TypeError("loop_fn must be a callable")
parallel_iterations = parallel_iterations or 32
# Create a new scope in which the caching device is either
# determined by the parent scope, or is set to place the cached
# Variable using the same placement as for the rest of the RNN.
with vs.variable_scope(scope or "rnn") as varscope:
if varscope.caching_device is None:
varscope.set_caching_device(lambda op: op.device)
time = constant_op.constant(0, dtype=dtypes.int32)
(elements_finished, next_input, initial_state, emit_structure,
init_loop_state) = loop_fn(
time, None, None, None) # time, cell_output, cell_state, loop_state
flat_input = nest.flatten(next_input)
# Need a surrogate loop state for the while_loop if none is available.
loop_state = (init_loop_state if init_loop_state is not None
else constant_op.constant(0, dtype=dtypes.int32))
input_shape = [input_.get_shape() for input_ in flat_input]
static_batch_size = input_shape[0][0]
for input_shape_i in input_shape:
# Static verification that batch sizes all match
static_batch_size.merge_with(input_shape_i[0])
batch_size = static_batch_size.value
if batch_size is None:
batch_size = array_ops.shape(flat_input[0])[0]
nest.assert_same_structure(initial_state, cell.state_size)
state = initial_state
flat_state = nest.flatten(state)
flat_state = [ops.convert_to_tensor(s) for s in flat_state]
state = nest.pack_sequence_as(structure=state,
flat_sequence=flat_state)
if emit_structure is not None:
flat_emit_structure = nest.flatten(emit_structure)
flat_emit_size = [emit.get_shape() for emit in flat_emit_structure]
flat_emit_dtypes = [emit.dtype for emit in flat_emit_structure]
else:
emit_structure = cell.output_size
flat_emit_size = nest.flatten(emit_structure)
flat_emit_dtypes = [flat_state[0].dtype] * len(flat_emit_size)
flat_emit_ta = [
tensor_array_ops.TensorArray(
dtype=dtype_i, dynamic_size=True, size=0, name="rnn_output_%d" % i)
for i, dtype_i in enumerate(flat_emit_dtypes)]
emit_ta = nest.pack_sequence_as(structure=emit_structure,
flat_sequence=flat_emit_ta)
flat_zero_emit = [
array_ops.zeros(
_state_size_with_prefix(size_i, prefix=[batch_size]),
dtype_i)
for size_i, dtype_i in zip(flat_emit_size, flat_emit_dtypes)]
zero_emit = nest.pack_sequence_as(structure=emit_structure,
flat_sequence=flat_zero_emit)
def condition(unused_time, elements_finished, *_):
return math_ops.logical_not(math_ops.reduce_all(elements_finished))
def body(time, elements_finished, current_input,
emit_ta, state, loop_state):
"""Internal while loop body for raw_rnn.
Args:
time: time scalar.
elements_finished: batch-size vector.
current_input: possibly nested tuple of input tensors.
emit_ta: possibly nested tuple of output TensorArrays.
state: possibly nested tuple of state tensors.
loop_state: possibly nested tuple of loop state tensors.
Returns:
Tuple having the same size as Args but with updated values.
"""
(next_output, cell_state) = cell(current_input, state)
nest.assert_same_structure(state, cell_state)
nest.assert_same_structure(cell.output_size, next_output)
next_time = time + 1
(next_finished, next_input, next_state, emit_output,
next_loop_state) = loop_fn(
next_time, next_output, cell_state, loop_state)
nest.assert_same_structure(state, next_state)
nest.assert_same_structure(current_input, next_input)
nest.assert_same_structure(emit_ta, emit_output)
# If loop_fn returns None for next_loop_state, just reuse the
# previous one.
loop_state = loop_state if next_loop_state is None else next_loop_state
def _copy_some_through(current, candidate):
"""Copy some tensors through via array_ops.where."""
current_flat = nest.flatten(current)
candidate_flat = nest.flatten(candidate)
# pylint: disable=g-long-lambda,cell-var-from-loop
result_flat = [
_on_device(
lambda: array_ops.where(
elements_finished, current_i, candidate_i),
device=candidate_i.op.device)
for (current_i, candidate_i) in zip(current_flat, candidate_flat)]
# pylint: enable=g-long-lambda,cell-var-from-loop
return nest.pack_sequence_as(
structure=current, flat_sequence=result_flat)
emit_output = _copy_some_through(zero_emit, emit_output)
next_state = _copy_some_through(state, next_state)
emit_output_flat = nest.flatten(emit_output)
emit_ta_flat = nest.flatten(emit_ta)
elements_finished = math_ops.logical_or(elements_finished, next_finished)
emit_ta_flat = [
ta.write(time, emit)
for (ta, emit) in zip(emit_ta_flat, emit_output_flat)]
emit_ta = nest.pack_sequence_as(
structure=emit_structure, flat_sequence=emit_ta_flat)
return (next_time, elements_finished, next_input,
emit_ta, next_state, loop_state)
returned = control_flow_ops.while_loop(
condition, body, loop_vars=[
time, elements_finished, next_input,
emit_ta, state, loop_state],
parallel_iterations=parallel_iterations,
swap_memory=swap_memory)
(emit_ta, final_state, final_loop_state) = returned[-3:]
if init_loop_state is None:
final_loop_state = None
return (emit_ta, final_state, final_loop_state)
| 42.787356 | 80 | 0.701903 | [
"Apache-2.0"
] | gameover27/hiptensorflow | tensorflow/python/ops/rnn.py | 44,670 | Python |
"""
Module containing NetPyNE metadata
"""
from __future__ import unicode_literals
from __future__ import print_function
from __future__ import division
from __future__ import absolute_import
from future import standard_library
standard_library.install_aliases()
metadata = {
# ---------------------------------------------------------------------------------------------------------------------
# netParams
# ---------------------------------------------------------------------------------------------------------------------
"netParams": {
"label": "Network Parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"popParams": {
"label": "Population Parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"cellType": {
"label": "Cell type",
"suggestions": "",
"help": "Arbitrary cell type attribute/tag assigned to all cells in this population; can be used as condition to apply specific cell properties. e.g. 'Pyr' (for pyramidal neurons) or 'FS' (for fast-spiking interneurons)",
"hintText": "",
"type": "str"
},
"numCells": {
"label": "Number of cells",
"suggestions": "",
"help": "The total number of cells in this population.",
"hintText": "number of cells",
"type": "int"
},
"density": {
"label": "Cell density (neurons/mm^3)",
"suggestions": "",
"help": "The cell density in neurons/mm3. The volume occupied by each population can be customized (see xRange, yRange and zRange); otherwise the full network volume will be used (defined in netParams: sizeX, sizeY, sizeZ). density can be expressed as a function of normalized location (xnorm, ynorm or znorm), by providing a string with the variable and any common Python mathematical operators/functions. e.g. '1e5 * exp(-ynorm/2)'. ",
"hintText": "density in neurons/mm3",
"type": "str"
},
"gridSpacing": {
"label": "Grid spacing (um)",
"suggestions": "",
"help": "Fixed grid spacing between cells (in um). Cells will be placed in a grid, with the total number of cells be determined based on spacing and sizeX, sizeY, sizeZ. e.g. a spacing of 20 with sizeX=sizeY=sizeZ=100 will lead to 5*5*5=125 cells.",
"hintText": "fixed grid spacing",
"type": "int"
},
"cellModel": {
"label": "Cell model",
"help": "Can be either 1) an arbitrary cell model attribute/tag assigned to all cells in this population, and used later as a condition to apply specific cell properties. e.g. 'HH' (standard Hodkgin-Huxley type cell model) or 'Izhi2007' (Izhikevich point neuron model), 2) a point process artificial cell, with its parameters defined directly in this population entry, i.e. no need for cell propoerties (e.g. 'NetStim', VecStim', 'IntFire1')",
"suggestions": [
"VecStim",
"NetStim",
"IntFire1"
],
"type": "str"
},
"xRange": {
"label": "X-axis range (um)",
"help": "Range of neuron positions in x-axis (horizontal length), specified as a 2-element list [min, max] using absolute values in um (e.g.[100, 200]).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"xnormRange": {
"label": "X-axis normalized range (0-1)",
"help": "Range of neuron positions in x-axis (horizontal length), specified as a 2-element list [min, max] using normalized values between 0 and 1 as fraction of sizeX (e.g.[0.1,0.2]).",
"suggestions": "",
"hintText": "",
"default": [
0,
1
],
"type": "list(float)"
},
"yRange": {
"label": "Y-axis range (um)",
"help": "Range of neuron positions in y-axis (vertical height=cortical depth), specified as 2-element list [min, max] using absolute values in um (e.g.[100,200]).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"ynormRange": {
"label": "Y-axis normalized range (0-1)",
"help": "Range of neuron positions in y-axis (vertical height=cortical depth), specified as a 2-element list [min, max] using normalized values between 0 and 1 as fraction of sizeY (e.g.[0.1,0.2]).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"zRange": {
"label": "Z-axis range (um)",
"help": "Range of neuron positions in z-axis (horizontal depth), specified as a 2-element list [min, max] using absolute value in um (e.g.[100,200]).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"znormRange": {
"label": "Z-axis normalized range (0-1)",
"help": "Range of neuron positions in z-axis (horizontal depth), specified as a 2-element list [min, max] using normalized values between 0 and 1 as fraction of sizeZ (e.g.[0.1,0.2]).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"interval": {
"label": "Spike interval (ms)",
"help": "Spike interval in ms.",
"suggestions": "",
"hintText": "50",
"type": "float"
},
"rate": {
"label": "Firing rate (Hz)",
"help": "Firing rate in Hz (note this is the inverse of the NetStim interval property).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"noise": {
"label": "Noise fraction (0-1)",
"help": "Fraction of noise in NetStim (0 = deterministic; 1 = completely random).",
"suggestions": "",
"hintText": "0.5",
"type": "list(float)"
},
"start": {
"label": "Start time (ms)",
"help": "Time of first spike in ms (default = 0).",
"suggestions": "",
"hintText": "0",
"type": "list(float)"
},
"number": {
"label": "Max number of spikes",
"help": "Max number of spikes generated (default = 1e12).",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"seed": {
"label": "Randomizer seed (optional)",
"help": " Seed for randomizer (optional; defaults to value set in simConfig.seeds['stim'])",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"spkTimes": {
"label": "Spike times",
"help": "List of spike times (only for 'VecStim') e.g. [1, 10, 40, 50], range(1,500,10), or any variable containing a Python list.",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"pulses": {
"label": "Pulses",
"help": "List of spiking pulses (only for 'VecStim'); each item includes the start (ms), end (ms), rate (Hz), and noise (0 to 1) pulse parameters. ",
"suggestions": "",
"hintText": "",
"type": "list(float)"
}
}
},
"scale": {
"label": "scale factor",
"help": "Scale factor multiplier for number of cells (default: 1)",
"suggestions": "",
"hintText": "",
"default": 1,
"type": "float"
},
"shape": {
"label": "network shape",
"help": "Shape of network: 'cuboid', 'cylinder' or 'ellipsoid' (default: 'cuboid')",
"suggestions": "",
"hintText": "",
"options": [
"cuboid",
"cylinder",
"ellipsoid"
],
"default": "cuboid",
"type": "str"
},
"sizeX": {
"label": "x-dimension",
"help": "x-dimension (horizontal length) network size in um (default: 100)",
"suggestions": "",
"hintText": "",
"default": 100,
"type": "float"
},
"sizeY": {
"label": "y-dimension",
"help": "y-dimension (horizontal length) network size in um (default: 100)",
"suggestions": "",
"hintText": "",
"default": 100,
"type": "float"
},
"sizeZ": {
"label": "z-dimension",
"help": "z-dimension (horizontal length) network size in um (default: 100)",
"suggestions": "",
"hintText": "",
"default": 100,
"type": "float"
},
"rotateCellsRandomly": {
"label": "random rotation",
"help": "Random rotation of cells around y-axis [min,max] radians, e.g. [0, 3.0] (default: False)",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"defaultWeight": {
"label": "default weight connection",
"help": "Default connection weight (default: 1)",
"suggestions": "",
"hintText": "",
"default": 1,
"type": "float"
},
"defaultDelay": {
"label": "default delay",
"help": "Default connection delay, in ms (default: 1)",
"suggestions": "",
"hintText": "",
"default": 1,
"type": "float"
},
"propVelocity": {
"label": "conduction velocity",
"help": "Conduction velocity in um/ms (e.g. 500 um/ms = 0.5 m/s) (default: 500)",
"suggestions": "",
"hintText": "",
"default": 500,
"type": "float"
},
"scaleConnWeight": {
"label": "connection weight scale factor",
"help": "Connection weight scale factor (excludes NetStims) (default: 1)",
"suggestions": "",
"hintText": "",
"default": 1,
"type": "float"
},
"scaleConnWeightNetStims": {
"label": "connection weight scale factor for NetStims",
"help": "Connection weight scale factor for NetStims (default: 1)",
"suggestions": "",
"hintText": "",
"default": 1,
"type": "float"
},
"scaleConnWeightModels": {
"label": "Connection weight scale factor for each cell model",
"help": "Connection weight scale factor for each cell model, e.g. {'HH': 0.1, 'Izhi': 0.2} (default: {})",
"suggestions": "",
"hintText": "",
"type": "dict"
},
"popTagsCopiedToCells": {
"label": "",
"help": "List of tags that will be copied from the population to the cells (default: ['pop', 'cellModel', 'cellType'])}",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.cellParams
# ---------------------------------------------------------------------------------------------------------------------
"cellParams": {
"label": "Cell Parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"conds": {
"label": "Conds",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"pop": {
"label": "Population",
"help": "Apply the cell rule only to cells belonging to this population (or list of populations).",
"suggestions": "",
"hintText": "",
"type": "list(str)"
},
"cellType": {
"label": "Cell type",
"suggestions": "",
"help": "Apply the cell rule only to cells with this cell type attribute/tag.",
"hintText": "",
"type": "list(str)"
},
"cellModel": {
"label": "Cell model",
"suggestions": "",
"help": "Apply the cell rule only to cells with this cell model attribute/tag.",
"hintText": "",
"type": "list(str)"
},
"x": {
"label": "Range of x-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these x-axis locations.",
"hintText": ""
},
"y": {
"label": "Range of y-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these y-axis locations.",
"hintText": ""
},
"z": {
"label": "Range of z-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these z-axis locations.",
"hintText": ""
},
"xnorm": {
"label": "Range of normalized x-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these normalized x-axis locations.",
"hintText": ""
},
"ynorm": {
"label": "Range of normalized y-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these normalized y-axis locations.",
"hintText": ""
},
"znorm": {
"label": "Range of normalized z-axis locations",
"suggestions": "",
"help": "Apply the cell rule only to cells within these normalized z-axis locations.",
"hintText": ""
}
}
},
"secs": {
"label": "Sections",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"geom": {
"label": "Cell geometry",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"diam": {
"label": "Diameter (um)",
"default": 10,
"suggestions": "",
"help": "",
"hintText": "10",
"type": "float"
},
"L": {
"label": "Length (um)",
"default": 50,
"suggestions": "",
"help": "",
"hintText": "50",
"type": "float"
},
"Ra": {
"label": "Axial resistance, Ra (ohm-cm)",
"default": 100,
"suggestions": "",
"help": "",
"hintText": "100",
"type": "float"
},
"cm": {
"label": "Membrane capacitance, cm (uF/cm2)",
"suggestions": "",
"help": "",
"hintText": "1",
"type": "float"
},
"pt3d": {
"label": "3D points",
"suggestions": "",
"help": "",
"hintText": "",
"type": "list(list(float))"
},
"nseg": {
"label": "Number of segments, nseg",
"default": 1,
"suggestions": "",
"help": "",
"hintText": "1",
"type": "float"
}
},
"mechs": {
"label": "Mechanisms",
"help": "Dictionary of density/distributed mechanisms, including the name of the mechanism (e.g. hh or pas) and a list of properties of the mechanism (e.g. {'g': 0.003, 'e': -70}).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"ions": {
"label": "Ions",
"help": "Dictionary of ions, including the name of the ion (e.g. hh or pas) and a list of properties of the ion (e.g. {'e': -70}).",
"suggestions": "",
"hintText": ""
},
"pointps": {
"label": "Point processes",
"help": "Dictionary of point processes (excluding synaptic mechanisms). The key contains an arbitrary label (e.g. 'Izhi') The value contains a dictionary with the point process properties (e.g. {'mod':'Izhi2007a', 'a':0.03, 'b':-2, 'c':-50, 'd':100, 'celltype':1}).",
"suggestions": "",
"hintText": "",
"children": {
"mod": {
"label": "Point process name",
"help": "The name of the NEURON mechanism, e.g. 'Izhi2007a'",
"suggestions": "",
"hintText": "",
"type": "float"
},
"loc": {
"label": "Location (0-1)",
"help": "Section location where to place synaptic mechanism, e.g. 1.0, default=0.5.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"vref": {
"label": "Point process variable for voltage (optional)",
"help": "Internal mechanism variable containing the cell membrane voltage, e.g. 'V'.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"synList": {
"label": "Point process list of synapses (optional)",
"help": "list of internal mechanism synaptic mechanism labels, e.g. ['AMPA', 'NMDA', 'GABAB'].",
"suggestions": "",
"hintText": "",
"type": "float"
}
},
"vinit": {
"label": "Initial membrance voltage, vinit (mV)",
"help": "(optional) Initial membrane voltage (in mV) of the section (default: -65).e.g. cellRule['secs']['soma']['vinit'] = -72",
"suggestions": "",
"hintText": ""
},
"spikeGenLoc": {
"label": "Spike generation location (0-1)",
"help": "(optional) Indicates that this section is responsible for spike generation (instead of the default 'soma'), and provides the location (segment) where spikes are generated.e.g. cellRule['secs']['axon']['spikeGenLoc'] = 1.0.",
"suggestions": "",
"hintText": ""
},
"threshold": {
"label": "Spike threshold voltage (mV)",
"help": "(optional) Threshold voltage (in mV) used to detect a spike originating in this section of the cell. If omitted, defaults to netParams.defaultThreshold = 10.0.e.g. cellRule['secs']['soma']['threshold'] = 5.0.",
"suggestions": "",
"hintText": ""
}
},
"secLists": {
"label": "Section lists (optional) ",
"help": "Dictionary of sections lists (e.g. {'all': ['soma', 'dend']})",
"suggestions": "",
"hintText": ""
}
},
"topol": {
"label": "Topology",
"help": "Topological properties, including parentSec (label of parent section), parentX (parent location where to make connection) and childX (current section child location where to make connection).",
"suggestions": "",
"hintText": "",
"children": {
"parentSec": {
"label": "Parent Section",
"suggestions": [
"soma"
],
"help": "label of parent section",
"hintText": "soma",
"type": "str"
},
"parentX": {
"label": "Parent connection location",
"suggestions": [
0,
1
],
"help": "Parent location where to make connection",
"hintText": "1",
"type": "float"
},
"childX": {
"label": "Child connection location",
"suggestions": [
0,
1
],
"help": "Current section child location where to make connection",
"hintText": "1",
"type": "float"
}
}
}
}
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.synMechParams
# ---------------------------------------------------------------------------------------------------------------------
"synMechParams": {
"label": "Synaptic mechanism parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"mod": {
"label": "NMODL mechanism name",
"help": "The NMODL mechanism name (e.g. 'ExpSyn'); note this does not always coincide with the name of the mod file.",
"suggestions": "",
"options": [
"ExpSyn",
"Exp2Syn"
],
"hintText": "",
"type": "str"
},
"selfNetCon": {
"label": "Self NetCon parameters",
"help": "Dict with parameters of NetCon between the cell voltage and the synapse, required by some synaptic mechanisms such as the homeostatic synapse (hsyn). e.g. 'selfNetCon': {'sec': 'soma' , threshold: -15, 'weight': -1, 'delay': 0} (by default the source section, 'sec' = 'soma').",
"suggestions": "",
"hintText": ""
},
"tau1": {
"label": "Time constant for exponential 1 (ms)",
"help": "Define the time constant for the first exponential.",
"suggestions": "",
"hintText": "1",
"type": "float"
},
"tau2": {
"label": "Time constant for exponential 2 (ms)",
"help": "Define the time constant for the second exponential.",
"suggestions": "",
"hintText": "5",
"type": "float"
},
"e": {
"label": "Reversal potential (mV)",
"help": "Reversal potential of the synaptic receptors.",
"suggestions": "",
"hintText": "0",
"type": "float"
},
"i": {
"label": "synaptic current (nA)",
"help": "Synaptic current in nA.",
"suggestions": "",
"hintText": "10",
"type": "float"
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.connParams
# ---------------------------------------------------------------------------------------------------------------------
"connParams": {
"label": "Connectivity parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"preConds": {
"label": "Conditions for the presynaptic cells",
"help": "Presynaptic cell conditions defined using attributes/tags and the required value e.g. {'cellType': 'PYR'}. Values can be lists, e.g. {'pop': ['Exc1', 'Exc2']}. For location properties, the list values correspond to the min and max values, e.g. {'ynorm': [0.1, 0.6]}.",
"suggestions": "",
"hintText": "",
"children": {
"pop": {
"label": "Population (multiple selection available)",
"suggestions": "",
"help": "Cells belonging to this population (or list of populations) will be connected pre-synaptically.",
"hintText": ""
},
"cellType": {
"label": "Cell type (multiple selection available)",
"suggestions": "",
"help": "Ccells with this cell type attribute/tag will be connected pre-synaptically.",
"hintText": ""
},
"cellModel": {
"label": "Cell model (multiple selection available)",
"suggestions": "",
"help": "Cells with this cell model attribute/tag will be connected pre-synaptically.",
"hintText": ""
},
"x": {
"label": "Range of x-axis locations",
"suggestions": "",
"help": "Cells within these x-axis locations will be connected pre-synaptically.",
"hintText": ""
},
"y": {
"label": "Range of y-axis locations",
"suggestions": "",
"help": "Cells within these y-axis locations will be connected pre-synaptically.",
"hintText": ""
},
"z": {
"label": "Range of z-axis locations",
"suggestions": "",
"help": "Cells within these z-axis locations will be connected pre-synaptically..",
"hintText": ""
},
"xnorm": {
"label": "Range of normalized x-axis locations",
"suggestions": "",
"help": "Cells within these normalized x-axis locations will be connected pre-synaptically.",
"hintText": ""
},
"ynorm": {
"label": "Range of normalized y-axis locations",
"suggestions": "",
"help": "Cells within these normalized y-axis locations will be connected pre-synaptically.",
"hintText": ""
},
"znorm": {
"label": "Range of normalized z-axis locations",
"suggestions": "",
"help": "Cells within these normalized z-axis locations will be connected pre-synaptically.",
"hintText": ""
}
}
},
"postConds": {
"label": "Conditions for the postsynaptic cells",
"help": "Defined as a dictionary with the attributes/tags of the postsynaptic cell and the required values e.g. {'cellType': 'PYR'}. Values can be lists, e.g. {'pop': ['Exc1', 'Exc2']}. For location properties, the list values correspond to the min and max values, e.g. {'ynorm': [0.1, 0.6]}.",
"suggestions": "",
"hintText": "",
"children": {
"pop": {
"label": "Population (multiple selection available)",
"suggestions": "",
"help": "Cells belonging to this population (or list of populations) will be connected post-synaptically.",
"hintText": ""
},
"cellType": {
"label": "Cell type (multiple selection available)",
"suggestions": "",
"help": "Ccells with this cell type attribute/tag will be connected post-synaptically.",
"hintText": ""
},
"cellModel": {
"label": "Cell model (multiple selection available)",
"suggestions": "",
"help": "Cells with this cell model attribute/tag will be connected post-synaptically.",
"hintText": ""
},
"x": {
"label": "Range of x-axis locations",
"suggestions": "",
"help": "Cells within these x-axis locations will be connected post-synaptically.",
"hintText": ""
},
"y": {
"label": "Range of y-axis locations",
"suggestions": "",
"help": "Cells within these y-axis locations will be connected post-synaptically.",
"hintText": ""
},
"z": {
"label": "Range of z-axis locations",
"suggestions": "",
"help": "Cells within these z-axis locations will be connected post-synaptically..",
"hintText": ""
},
"xnorm": {
"label": "Range of normalized x-axis locations",
"suggestions": "",
"help": "Cells within these normalized x-axis locations will be connected post-synaptically.",
"hintText": ""
},
"ynorm": {
"label": "Range of normalized y-axis locations",
"suggestions": "",
"help": "Cells within these normalized y-axis locations will be connected post-synaptically.",
"hintText": ""
},
"znorm": {
"label": "Range of normalized z-axis locations",
"suggestions": "",
"help": "Cells within these normalized z-axis locations will be connected post-synaptically.",
"hintText": ""
}
}
},
"sec": {
"label": "Postsynaptic neuron section",
"help": "Name of target section on the postsynaptic neuron (e.g. 'soma'). If omitted, defaults to 'soma' if exists, otherwise to first section in the cell sections list. If synsPerConn > 1, a list of sections or sectionList can be specified, and synapses will be distributed uniformly along the specified section(s), taking into account the length of each section.",
"suggestions": "",
"hintText": "soma",
"type": "list(str)"
},
"loc": {
"label": "Postsynaptic neuron location (0-1)",
"help": "Location of target synaptic mechanism (e.g. 0.3). If omitted, defaults to 0.5. Can be single value, or list (if have synsPerConn > 1) or list of lists (If have both a list of synMechs and synsPerConn > 1).",
"suggestions": "",
"hintText": "0.5",
"type": "list(float)"
},
"synMech": {
"label": "Synaptic mechanism",
"help": "Label (or list of labels) of target synaptic mechanism on the postsynaptic neuron (e.g. 'AMPA' or ['AMPA', 'NMDA']). If omitted employs first synaptic mechanism in the cell synaptic mechanisms list. If have list, a separate connection is created to each synMech; and a list of weights, delays and or locs can be provided.",
"suggestions": "",
"hintText": ""
},
"synsPerConn": {
"label": "Number of individual synaptic contacts per connection",
"help": "Number of individual synaptic contacts (synapses) per cell-to-cell connection (connection). Can be defined as a function (see Functions as strings). If omitted, defaults to 1.",
"suggestions": "",
"hintText": "",
"default": 1
},
"weight": {
"label": "Weight of synaptic connection",
"help": "Strength of synaptic connection (e.g. 0.01). Associated to a change in conductance, but has different meaning and scale depending on the synaptic mechanism and cell model. Can be defined as a function (see Functions as strings). If omitted, defaults to netParams.defaultWeight = 1.",
"suggestions": "",
"hintText": "",
"type": "func"
},
"delay": {
"label": "Connection delay (ms)",
"help": "Time (in ms) for the presynaptic spike to reach the postsynaptic neuron. Can be defined as a function (see Functions as strings). If omitted, defaults to netParams.defaultDelay = 1.",
"suggestions": "",
"hintText": "",
"type": "func"
},
"probability": {
"label": "Probability of connection (0-1)",
"help": "Probability of connection between each pre and postsynaptic cell (0 to 1). Can be a string that defines as a function, e.g. '0.1*dist_3D+uniform(0.2,0.4)' (see Documentation on 'Functions as strings'). Overrides the convergence, divergence and fromList parameters.",
"suggestions": "0.1",
"hintText": "",
"type": "func"
},
"convergence": {
"label": "Convergence",
"help": "Number of pre-synaptic cells connected to each post-synaptic cell. Can be a string that defines as a function, e.g. '2*dist_3D+uniform(2,4)' (see Documentation on 'Functions as strings'). Overrides the divergence and fromList parameters.",
"suggestions": "5",
"hintText": "",
"type": "func"
},
"divergence": {
"label": "Divergence",
"help": "Number of post-synaptic cells connected to each pre-synaptic cell. Can be a string that defines as a function, e.g. '2*dist_3D+uniform(2,4)' (see Documentation on 'Functions as strings'). Overrides the fromList parameter.",
"suggestions": "5",
"hintText": "",
"type": "func"
},
"connList": {
"label": "Explicit list of one-to-one connections",
"help": "Each connection is indicated with relative ids of cell in pre and post populations, e.g. [[0,1],[3,1]] creates a connection between pre cell 0 and post cell 1; and pre cell 3 and post cell 1. Weights, delays and locs can also be specified as a list for each of the individual cell connection. These lists can be 2D or 3D if combined with multiple synMechs and synsPerConn > 1 (the outer dimension will correspond to the connList).",
"suggestions": "",
"hintText": "list(list(float))"
},
"connFunc": {
"label": "Internal connectivity function to use (not required)",
"help": "Automatically set to probConn, convConn, divConn or fromList, when the probability, convergence, divergence or connList parameters are included, respectively. Otherwise defaults to fullConn, ie. all-to-all connectivity.",
"suggestions": "",
"hintText": ""
},
"shape": {
"label": "Weight shape",
"help": "Modifies the conn weight dynamically during the simulation based on the specified pattern. Contains a dictionary with the following fields: 'switchOnOff' - times at which to switch on and off the weight, 'pulseType' - type of pulse to generate; either 'square' or 'gaussian', 'pulsePeriod' - period (in ms) of the pulse, 'pulseWidth' - width (in ms) of the pulse.",
"suggestions": "",
"hintText": ""
},
"plasticity": {
"label": "Plasticity mechanism",
"help": "Requires 2 fields: mech to specifiy the name of the plasticity mechanism, and params containing a dictionary with the parameters of the mechanism, e.g. {'mech': 'STDP', 'params': {'hebbwt': 0.01, 'antiwt':-0.01, 'wmax': 50, 'RLon': 1 'tauhebb': 10}}.",
"suggestions": "",
"hintText": "",
"type": "dict"
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.stimSourceParams
# ---------------------------------------------------------------------------------------------------------------------
"stimSourceParams": {
"label": "Stimulation source parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"type": {
"label": "Point process used as stimulator",
"help": "Point process used as stimulator; allowed values: 'IClamp', 'VClamp', 'SEClamp', 'NetStim' and 'AlphaSynapse'. Note that NetStims can be added both using this method, or by creating a population of 'cellModel': 'NetStim' and adding the appropriate connections.",
"suggestions": "",
"hintText": "",
"default": "IClamp",
"type": "str"
},
"dur": {
"label": "Current clamp duration (ms)",
"help": "Duration of current clamp injection in ms",
"suggestions": "",
"hintText": "10",
"type": "float"
},
"amp": {
"label": "Current clamp amplitude (nA)",
"help": "Amplitude of current injection in nA",
"suggestions": "",
"hintText": "10",
"type": "float"
},
"del": {
"label": "Current clamp delay (ms)",
"help": "Delay (time when turned on after simulation starts) of current clamp in ms.",
"suggestions": "",
"hintText": "5",
"type": "float"
},
"vClampAmp": {
"label": "Current clamp amplitude (nA)",
"help": "Voltage clamp with three levels. Clamp is on at time 0, and off at time dur[0]+dur[1]+dur[2].",
"suggestions": "",
"hintText": "10",
"type": "list(float)"
},
"vClampDur": {
"label": "Current clamp delay (ms)",
"help": "Voltage clamp with three levels. Clamp is on at time 0, and off at time dur[0]+dur[1]+dur[2].",
"suggestions": "",
"hintText": "5",
"type": "list(float)"
},
"interval": {
"label": "Interval between spikes (ms)",
"help": "Define the mean time interval between spike.",
"suggestions": "10",
"hintText": "",
"type": "float"
},
"rate": {
"label": "Firing rate (Hz)",
"help": "Firing rate in Hz (note this is the inverse of the NetStim interval property).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"rstim": {
"label": "Voltage clamp stimulation resistance",
"help": "Voltage clamp stimulation resistance.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"gain": {
"label": "Voltage clamp amplifier gain",
"help": "Voltage clamp amplifier gain.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"number": {
"label": "Maximum number of spikes",
"help": "Maximum number of spikes generated by the NetStim.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"start": {
"label": "Start time of first spike",
"help": "Define the start time for the first spike.",
"suggestions": "0",
"hintText": "",
"type": "float"
},
"noise": {
"label": "Noise/randomness fraction (0-1)",
"help": "Fractional noise, 0 <= noise <= 1, means that an interval between spikes consists of a fixed interval of duration (1 - noise)*interval plus a negexp interval of mean duration noise*interval. Note that the most likely negexp interval has duration 0.",
"suggestions": "0.5",
"hintText": "",
"type": "float"
},
"tau1": {
"label": "Voltage clamp tau1",
"help": "Voltage clamp tau1.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"tau2": {
"label": "Voltage clamp tau2",
"help": "Voltage clamp tau2.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"i": {
"label": "Voltage clamp current (nA)",
"help": "Voltage clamp injected current in nA.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"onset": {
"label": "Alpha synapse onset time (ms)",
"help": "Alpha synapse onset time.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"tau": {
"label": "Alpha synapse time constant (ms)",
"help": "Alpha synapse time constant (ms).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"gmax": {
"label": "Alpha synapse maximum conductance",
"help": "Alpha synapse maximum conductance.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"e": {
"label": "Alpha synapse equilibrium potential",
"help": "Alpha synapse equilibrium potential.",
"suggestions": "",
"hintText": "",
"type": "float"
},
"rs": {
"label": "Voltage clamp resistance (MOhm)",
"help": "Voltage clamp resistance (MOhm).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"vc": {
"label": "Voltage clamp reference voltage (mV)",
"help": "Voltage clamp reference voltage (mV).",
"suggestions": "",
"hintText": "",
"type": "float"
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.stimTargetParams
# ---------------------------------------------------------------------------------------------------------------------
"stimTargetParams": {
"label": "Stimulation target parameters",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"source": {
"label": "Stimulation source",
"help": "Label of the stimulation source (e.g. 'electrode_current').",
"suggestions": "",
"hintText": ""
},
"conds": {
"label": "Conditions of cells where the stimulation will be applied",
"help": "Conditions of cells where the stimulation will be applied. Can include a field 'cellList' with the relative cell indices within the subset of cells selected (e.g. 'conds': {'cellType':'PYR', 'y':[100,200], 'cellList': [1,2,3]}).",
"suggestions": "",
"hintText": "",
"children": {
"pop": {
"label": "Target population",
"help": "Populations that will receive the stimulation e.g. {'pop': ['Exc1', 'Exc2']}",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
"cellType": {
"label": "Target cell type",
"suggestions": "",
"help": "Cell types that will receive the stimulation",
"hintText": "",
"type": "str"
},
"cellModel": {
"label": "Target cell model",
"help": "Cell models that will receive the stimulation.",
"suggestions": "",
"type": "str"
},
"x": {
"label": "Range of x-axis locations",
"suggestions": "",
"help": "Cells within this x-axis locations will receive stimulation",
"hintText": ""
},
"y": {
"label": "Range of y-axis locations",
"suggestions": "",
"help": "Cells within this y-axis locations will receive stimulation",
"hintText": ""
},
"z": {
"label": "Range of z-axis locations",
"suggestions": "",
"help": "Cells within this z-axis locations will receive stimulation",
"hintText": ""
},
"xnorm": {
"label": "Range of normalized x-axis locations",
"suggestions": "",
"help": "Cells withing this normalized x-axis locations will receive stimulation",
"hintText": ""
},
"ynorm": {
"label": "Range of normalized y-axis locations",
"suggestions": "",
"help": "Cells within this normalized y-axis locations will receive stimulation",
"hintText": ""
},
"znorm": {
"label": "Range of normalized z-axis locations",
"suggestions": "",
"help": "Cells within this normalized z-axis locations will receive stimulation",
"hintText": ""
},
"cellList": {
"label": "Target cell global indices (gids)",
"help": "Global indices (gids) of neurons to receive stimulation. ([1, 8, 12])",
"suggestions": "",
"hintText": "",
"type": "list(float)"
},
}
},
"sec": {
"label": "Target section",
"help": "Target section (default: 'soma').",
"suggestions": "",
"hintText": "",
"type": "str"
},
"loc": {
"label": "Target location",
"help": "Target location (default: 0.5). Can be defined as a function (see Functions as strings).",
"suggestions": "",
"hintText": "",
"type": "float"
},
"synMech": {
"label": "Target synaptic mechanism",
"help": "Synaptic mechanism label to connect NetStim to. Optional; only for NetStims.",
"suggestions": "",
"hintText": ""
},
"weight": {
"label": "Weight of connection between NetStim and cell",
"help": "Weight of connection between NetStim and cell. Optional; only for NetStims. Can be defined as a function (see Functions as strings).",
"suggestions": "",
"hintText": ""
},
"delay": {
"label": "Delay of connection between NetStim and cell",
"help": "Delay of connection between NetStim and cell (default: 1). Optional; only for NetStims. Can be defined as a function (see Functions as strings).",
"suggestions": "",
"hintText": ""
},
"synsPerConn": {
"label": "Number of synaptic contacts per connection between NetStim and cell",
"help": "Number of synaptic contacts of connection between NetStim and cell (default: 1). Optional; only for NetStims. Can be defined as a function (see Functions as strings).",
"suggestions": "",
"hintText": ""
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# netParams.importCellParams
# ---------------------------------------------------------------------------------------------------------------------
"importCellParams": {
"label": "Import cell from .hoc or .py templates",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"fileName": {
"label": "Absolute path to file",
"help": "Absolute path to .hoc or .py template file.",
"suggestions": "",
"hintText": "",
"type": "str"
},
"cellName": {
"label": "Cell template/class name",
"help": "Template or class name defined inside the .hoc or .py file",
"suggestions": "",
"hintText": "",
"type": "str"
},
"label": {
"label": "Cell rule label",
"help": "Give a name to this cell rule.",
"suggestions": "",
"hintText": "",
"type": "str"
},
"importSynMechs": {
"label": "Import synaptic mechanisms",
"help": "If true, synaptic mechanisms will also be imported from the file. (default: False)",
"suggestions": "",
"hintText": "",
"type": "bool"
},
"compileMod": {
"label": "Compile mod files",
"help": "If true, mod files will be compiled before importing the cell. (default: false)",
"suggestions": "",
"hintText": "",
"type": "bool"
},
"modFolder": {
"label": "Path to mod folder",
"help": "Define the absolute path to the folder containing the mod files.",
"suggestions": "",
"hintText": "",
"type": "str"
},
}
}
}
},
# ---------------------------------------------------------------------------------------------------------------------
# simConfig
# ---------------------------------------------------------------------------------------------------------------------
"simConfig": {
"label": "Simulation Configuration",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"simLabel": {
"label": "Simulation label",
"help": "Choose a label for this simulation",
"suggestions": "",
"type": "str"
},
"duration": {
"label": "Duration (ms)",
"help": "Simulation duration in ms (default: 1000)",
"suggestions": "",
"default": 1000,
"type": "float"
},
"dt": {
"label": "Time step, dt",
"help": "Simulation time step in ms (default: 0.1)",
"suggestions": "",
"default": 0.025,
"type": "float"
},
"seeds": {
"label": "Randomizer seeds",
"help": "Dictionary with random seeds for connectivity, input stimulation, and cell locations (default: {'conn': 1, 'stim': 1, 'loc': 1}).",
"suggestions": "",
"type": "dict"
},
"addSynMechs": {
"label": "Add synaptic mechanisms",
"help": "Whether to add synaptic mechanisms or not (default: True).",
"suggestions": "",
"type": "bool"
},
"includeParamsLabel": {
"label": "Include parameter rule label",
"help": "Include label of parameters rule that created that cell, conn or stim (default: True).",
"suggestions": "",
"type": "bool"
},
"timing": {
"label": "Show timing",
"help": "Show and record timing of each process (default: True).",
"suggestions": "",
"type": "bool"
},
"verbose": {
"label": "Verbose mode",
"help": "Show detailed messages (default: False).",
"suggestions": "",
"type": "bool"
},
"saveFolder": {
"label": "Output folder",
"help": "Path where to save output data (default: '')",
"suggestions": "",
"type": "str"
},
"filename": {
"label": "Output file name",
"help": "Name of file to save model output (default: 'model_output')",
"suggestions": "",
"default": "model_output",
"type": "str"
},
"saveDataInclude": {
"label": "Data to include in output file",
"help": "Data structures to save to file (default: ['netParams', 'netCells', 'netPops', 'simConfig', 'simData'])",
"suggestions": "",
"type": "list(str)"
},
"timestampFilename": {
"label": "Add timestamp to file name",
"help": "Add timestamp to filename to avoid overwriting (default: False)",
"suggestions": "",
"type": "bool"
},
"savePickle": {
"label": "Save as Pickle",
"help": "Save data to pickle file (default: False).",
"suggestions": "",
"type": "bool"
},
"saveJson": {
"label": "Save as JSON",
"help": "Save dat to json file (default: False).",
"suggestions": "",
"type": "bool"
},
"saveMat": {
"label": "Save as MAT",
"help": "Save data to mat file (default: False).",
"suggestions": "",
"type": "bool"
},
"saveHDF5": {
"label": "Save as HDF5",
"help": "Save data to save to HDF5 file (under development) (default: False).",
"suggestions": "",
"type": "bool"
},
"saveDpk": {
"label": "Save as DPK",
"help": "Save data to .dpk pickled file (default: False).",
"suggestions": "",
"type": "bool"
},
"checkErrors": {
"label": "Check parameter errors",
"help": "check for errors (default: False).",
"suggestions": "",
"type": "bool"
},
"checkErrorsVerbose": {
"label": "Check parameter errors verbose mode",
"help": "check errors vervose (default: False)",
"suggestions": "",
"type": "bool"
},
"backupCfgFile": {
"label": "Copy simulation configuration file to this folder:",
"help": "Copy cfg file to folder, eg. ['cfg.py', 'backupcfg/'] (default: []).",
"suggestions": "",
"type": "list(str)"
},
"recordCells": {
"label": "Cells to record traces from",
"help": "List of cells from which to record traces. Can include cell gids (e.g. 5), population labels (e.g. 'S' to record from one cell of the 'S' population), or 'all', to record from all cells. NOTE: All cells selected in the include argument of simConfig.analysis['plotTraces'] will be automatically included in recordCells. (default: []).",
"suggestions": "",
"type": "list(float)"
},
"recordTraces": {
"label": "Traces to record from cells",
"help": "Dict of traces to record (default: {} ; example: {'V_soma': {'sec':'soma','loc':0.5,'var':'v'} }).",
"suggestions": "",
"type": "dict(dict)",
"default": "{\"V_soma\": {\"sec\": \"soma\", \"loc\": 0.5, \"var\": \"v\"}}"
},
"saveCSV": {
"label": "Save as CSV",
"help": "save cvs file (under development) (default: False)",
"suggestions": "",
"type": "bool"
},
"saveDat": {
"label": "Save as DAT ",
"help": "save .dat file (default: False)",
"suggestions": "",
"type": "bool"
},
"saveCellSecs": {
"label": "Store cell sections after simulation",
"help": "Save cell sections after gathering data from nodes post simulation; set to False to reduce memory required (default: True)",
"suggestions": "",
"type": "bool"
},
"saveCellConns": {
"label": "Store cell connections after simulation",
"help": "Save cell connections after gathering data from nodes post simulation; set to False to reduce memory required (default: True)",
"suggestions": "",
"type": "bool"
},
"recordStim": {
"label": "Record spikes of artificial stimulators (NetStims and VecStims)",
"help": "Record spikes of NetStims and VecStims (default: False).",
"suggestions": "",
"type": "bool"
},
"recordLFP": {
"label": "Record LFP electrode locations",
"help": "3D locations of local field potential (LFP) electrodes, e.g. [[50, 100, 50], [50, 200]] (default: False).",
"suggestions": "",
"type": "list(list(float))"
},
"saveLFPCells": {
"label": "Store LFP of individual cells",
"help": "Store LFP generated individually by each cell in sim.allSimData['LFPCells'].",
"suggestions": "",
"type": "bool"
},
"recordStep": {
"label": "Time step for data recording (ms)",
"help": "Step size in ms for data recording (default: 0.1).",
"suggestions": "",
"default": 0.1,
"type": "float"
},
"printRunTime": {
"label": "Interval to print run time at (s)",
"help": "Print run time at interval (in sec) specified here (eg. 0.1) (default: False).",
"suggestions": "",
"type": "float"
},
"printSynsAfterRule": {
"label": "Print total connections",
"help": "Print total connections after each conn rule is applied.",
"suggestions": "",
"type": "bool"
},
"printPopAvgRates": {
"label": "Print population average firing rates",
"help": "Print population avg firing rates after run (default: False).",
"suggestions": "",
"type": "bool"
},
"connRandomSecFromList": {
"label": "Select random sections from list for connection",
"help": "Select random section (and location) from list even when synsPerConn=1 (default: True).",
"suggestions": "",
"type": "bool"
},
"compactConnFormat": {
"label": "Use compact connection format (list instead of dicT)",
"help": "Replace dict format with compact list format for conns (need to provide list of keys to include) (default: False).",
"suggestions": "",
"type": "bool"
},
"gatherOnlySimData": {
"label": "Gather only simulation output data",
"help": "Omits gathering of net and cell data thus reducing gatherData time (default: False).",
"suggestions": "",
"type": "bool"
},
"createPyStruct": {
"label": "Create Python structure",
"help": "Create Python structure (simulator-independent) when instantiating network (default: True).",
"suggestions": "",
"type": "bool"
},
"createNEURONObj": {
"label": "Create NEURON objects",
"help": "Create runnable network in NEURON when instantiating netpyne network metadata (default: True).",
"suggestions": "",
"type": "bool"
},
"cvode_active": {
"label": "use CVode",
"help": "Use CVode variable time step (default: False).",
"suggestions": "",
"type": "bool"
},
"cache_efficient": {
"label": "use CVode cache_efficient",
"help": "Use CVode cache_efficient option to optimize load when running on many cores (default: False).",
"suggestions": "",
"type": "bool"
},
"hParams": {
"label": "Set global parameters (temperature, initial voltage, etc)",
"help": "Dictionary with parameters of h module (default: {'celsius': 6.3, 'v_init': -65.0, 'clamp_resist': 0.001}).",
"suggestions": "",
"type": "dict"
},
"saveTxt": {
"label": "Save as TXT",
"help": "Save data to txt file (under development) (default: False)",
"suggestions": "",
"type": "bool"
},
"saveTiming": {
"label": "Save timing data to file",
"help": " Save timing data to pickle file (default: False).",
"suggestions": "",
"type": "bool"
},
# ---------------------------------------------------------------------------------------------------------------------
# simConfig.analysis
# ---------------------------------------------------------------------------------------------------------------------
"analysis": {
"label": "Analysis",
"suggestions": "",
"help": "",
"hintText": "",
"children": {
"plotRaster": {
"label": "Raster plot",
"suggestions": "",
"help": "Plot raster (spikes over time) of network cells.",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "str"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range of spikes shown; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"maxSpikes": {
"label": "Maximum number of spikes to plot",
"suggestions": "",
"help": "maximum number of spikes that will be plotted (int).",
"hintText": "",
"type": "float"
},
"orderBy": {
"label": "Order by",
"suggestions": "",
"help": "Unique numeric cell property to order y-axis by, e.g. 'gid', 'ynorm', 'y' ('gid'|'y'|'ynorm'|...)",
"hintText": "",
"options": [
"gid",
"y",
"ynorm"
],
"type": "str"
},
"orderInverse": {
"label": "Invert y-axis",
"suggestions": "",
"help": "Invert the y-axis order (True|False)",
"hintText": "",
"type": "bool"
},
"labels": {
"label": "Population labels",
"suggestions": "",
"help": "Show population labels in a legend or overlayed on one side of raster ('legend'|'overlay'))",
"hintText": "",
"type": "str"
},
"popRates": {
"label": "Include population rates",
"suggestions": "",
"help": "Include population rates ('legend'|'overlay')",
"hintText": "",
"options": [
"legend",
"overlay"
],
"type": "str"
},
"spikeHist": {
"label": "Overlay spike histogram",
"suggestions": "",
"help": "overlay line over raster showing spike histogram (spikes/bin) (None|'overlay'|'subplot')",
"hintText": "",
"options": [
"None",
"overlay",
"subplot"
],
"type": "str"
},
"spikeHistBin": {
"label": "Bin size for histogram",
"suggestions": "",
"help": "Size of bin in ms to use for histogram (int)",
"hintText": "",
"type": "float"
},
"syncLines": {
"label": "Synchronization lines",
"suggestions": "",
"help": "calculate synchorny measure and plot vertical lines for each spike to evidence synchrony (True|False)",
"hintText": "",
"type": "bool"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": "str"
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotSpikeHist": {
"label": "Plot Spike Histogram",
"suggestions": "",
"help": "Plot spike histogram.",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range of spikes shown; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"binSize": {
"label": "bin size for histogram",
"suggestions": "",
"help": "Size of bin in ms to use for histogram (int)",
"hintText": "",
"type": "int"
},
"overlay": {
"label": "show overlay",
"suggestions": "",
"help": "Whether to overlay the data lines or plot in separate subplots (True|False)",
"hintText": "",
"type": "bool"
},
"graphType": {
"label": "type of Graph",
"suggestions": "",
"help": " Type of graph to use (line graph or bar plot) ('line'|'bar')",
"hintText": "",
"options": [
"line",
"bar"
],
"type": "str"
},
"yaxis": {
"label": "axis units",
"suggestions": "",
"help": "Units of y axis (firing rate in Hz, or spike count) ('rate'|'count')",
"hintText": "",
"options": [
"rate",
"count"
],
"type": "str"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotRatePSD": {
"label": "Plot Rate PSD",
"suggestions": "",
"help": "Plot spikes power spectral density (PSD).",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range of spikes shown; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"binSize": {
"label": "Bin size",
"suggestions": "",
"help": "Size of bin in ms to use (int)",
"hintText": "",
"type": "float"
},
"maxFreq": {
"label": "maximum frequency",
"suggestions": "",
"help": " Maximum frequency to show in plot (float).",
"hintText": "",
"type": "float"
},
"NFFT": {
"label": "Number of point",
"suggestions": "",
"help": "The number of data points used in each block for the FFT (power of 2)",
"hintText": "",
"type": "float"
},
"noverlap": {
"label": "Number of overlap points",
"suggestions": "",
"help": "Number of points of overlap between segments (< nperseg).",
"hintText": "",
"type": "float"
},
"smooth": {
"label": "Window size",
"suggestions": "",
"help": "Window size for smoothing; no smoothing if 0.",
"hintText": "",
"type": "float"
},
"overlay": {
"label": "Overlay data",
"suggestions": "",
"help": "Whether to overlay the data lines or plot in separate subplots (True|False).",
"hintText": "",
"type": "bool"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotSpikeStats": {
"label": "Plot Spike Statistics",
"suggestions": "",
"help": "Plot spike histogram.",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range of spikes shown; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"graphType": {
"label": "type of graph",
"suggestions": "",
"help": "Type of graph to use ('boxplot').",
"hintText": "",
"options": [
"boxplot"
],
"type": "str"
},
"stats": {
"label": "meassure type to calculate stats",
"suggestions": "",
"help": "List of types measure to calculate stats over: cell firing rates, interspike interval coefficient of variation (ISI CV), pairwise synchrony, and/or overall synchrony (sync measures calculated using PySpike SPIKE-Synchrony measure) (['rate', |'isicv'| 'pairsync' |'sync'|]).",
"hintText": "",
"options": [
"rate",
"isicv",
"pairsync",
"sync"
],
"type": "str"
},
"popColors": {
"label": "color for each population",
"suggestions": "",
"help": "Dictionary with color (value) used for each population/key.",
"hintText": "",
"type": "dict"
},
"figSize": {
"label": "figure size",
"suggestions": "",
"help": "Size of figure ((width, height)).",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotTraces": {
"label": "Plot Traces",
"suggestions": "",
"help": "Plot recorded traces (specified in simConfig.recordTraces).",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list(float)"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range for shown Traces ; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"overlay": {
"label": "overlay data",
"suggestions": "",
"help": "Whether to overlay the data lines or plot in separate subplots (True|False).",
"hintText": "",
"type": "bool"
},
"oneFigPer": {
"label": "plot one figure per cell/trace",
"suggestions": "",
"help": "Whether to plot one figure per cell or per trace (showing multiple cells) ('cell'|'trace').",
"hintText": "",
"options": [
"cell",
"traces"
],
"type": "str"
},
"rerun": {
"label": "re-run simulation",
"suggestions": "",
"help": "rerun simulation so new set of cells gets recorded (True|False).",
"hintText": "",
"type": "bool"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotLFP": {
"label": "Plot LFP",
"suggestions": "",
"help": "Plot LFP / extracellular electrode recordings (time-resolved, power spectral density, time-frequency and 3D locations).",
"hintText": "",
"children": {
"electrodes": {
"label": "electrode to show",
"suggestions": "",
"help": " List of electrodes to include; 'avg'=avg of all electrodes; 'all'=each electrode separately (['avg', 'all', 0, 1, ...]).",
"hintText": "",
"type": "list"
},
"plots": {
"label": "Select plot types to show (multiple selection available)",
"suggestions": "",
"help": "list of plot types to show (['timeSeries', 'PSD', 'timeFreq', 'locations']).",
"hintText": "",
"options": [
"timeSeries",
"PSD",
"spectrogram",
"locations"
],
"type": "str"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Time range for shown Traces ; if None shows all ([start,stop])",
"hintText": "",
"type": "list(float)"
},
"NFFT": {
"label": "NFFT",
"suggestions": "",
"help": "The number of data points used in each block for the FFT (power of 2) (float)",
"hintText": "",
"type": "float"
},
"noverlap": {
"label": "Overlap",
"suggestions": "",
"help": "Number of points of overlap between segments (int, < nperseg).",
"hintText": "",
"type": "float"
},
"maxFreq": {
"label": "Maximum Frequency",
"suggestions": "",
"help": "Maximum frequency shown in plot for PSD and time-freq (float).",
"hintText": "",
"type": "float"
},
"nperseg": {
"label": "Segment length (nperseg)",
"suggestions": "",
"help": "Length of each segment for time-freq (int).",
"hintText": "",
"type": "float"
},
"smooth": {
"label": "Window size",
"suggestions": "",
"help": "Window size for smoothing; no smoothing if 0 (int).",
"hintText": "",
"type": "float"
},
"separation": {
"label": "Separation factor",
"suggestions": "",
"help": "Separation factor between time-resolved LFP plots; multiplied by max LFP value (float).",
"hintText": "",
"type": "float"
},
"includeAxon": {
"label": "Include axon",
"suggestions": "",
"help": "Whether to show the axon in the location plot (boolean).",
"hintText": "",
"type": "bool"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotShape": {
"label": "Plot Shape",
"suggestions": "",
"help": "",
"hintText": "Plot 3D cell shape using Matplotlib or NEURON Interviews PlotShape.",
"children": {
"includePre": {
"label": "population (or cell by index) to presyn",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list"
},
"includePost": {
"label": "population (or cell by index) to postsyn",
"suggestions": "",
"help": "List of cells to include (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])])",
"hintText": "",
"type": "list"
},
"synStyle": {
"label": "synaptic marker style",
"suggestions": "",
"help": "Style of marker to show synapses (Matplotlib markers).",
"hintText": "",
"type": "str"
},
"dist": {
"label": "3D distance",
"suggestions": "",
"help": "3D distance (like zoom).",
"hintText": "",
"type": "float"
},
"synSize": {
"label": "synapses marker size",
"suggestions": "",
"help": "Size of marker to show synapses.",
"hintText": "",
"type": "float"
},
"cvar": {
"label": "variable to represent in shape plot",
"suggestions": "",
"help": "Variable to represent in shape plot ('numSyns'|'weightNorm').",
"hintText": "",
"options": [
"numSyns",
"weightNorm"
],
"type": "str"
},
"cvals": {
"label": "value to represent in shape plot",
"suggestions": "",
"help": "List of values to represent in shape plot; must be same as num segments (list of size num segments; ).",
"hintText": "",
"type": "list(float)"
},
"iv": {
"label": "use NEURON iv",
"suggestions": "",
"help": "Use NEURON Interviews (instead of matplotlib) to show shape plot (True|False).",
"hintText": "",
"type": "bool"
},
"ivprops": {
"label": "properties for iv",
"suggestions": "",
"help": "Dict of properties to plot using Interviews (dict).",
"hintText": "",
"type": "dict"
},
"showSyns": {
"label": "show synaptic connections in 3D",
"suggestions": "",
"help": "Show synaptic connections in 3D (True|False).",
"hintText": "",
"type": "bool"
},
"bkgColor": {
"label": "background color",
"suggestions": "",
"help": "RGBA list/tuple with bakcground color eg. (0.5, 0.2, 0.1, 1.0) (list/tuple with 4 floats).",
"hintText": "",
"type": "list(float)"
},
"showElectrodes": {
"label": "show electrodes",
"suggestions": "",
"help": "Show electrodes in 3D (True|False).",
"hintText": "",
"type": "bool"
},
"includeAxon": {
"label": "include Axon in shape plot",
"suggestions": "",
"help": "Include axon in shape plot (True|False).",
"hintText": "",
"type": "bool"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plot2Dnet": {
"label": "Plot 2D net",
"suggestions": "",
"help": "Plot 2D representation of network cell positions and connections.",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to show (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"showConns": {
"label": "show connections",
"suggestions": "",
"help": "Whether to show connections or not (True|False).",
"hintText": "",
"type": "bool"
},
"view": {
"label": "perspective view",
"suggestions": "",
"help": "Perspective view, either front ('xy') or top-down ('xz').",
"hintText": "",
"options": [
"xy",
"xz"
],
"type": "str"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"plotConn": {
"label": "Plot Connectivity",
"suggestions": "",
"help": "Plot network connectivity.",
"hintText": "",
"children": {
"include": {
"label": "Cells to include",
"suggestions": "",
"help": "List of cells to show (['all'|,'allCells'|,'allNetStims'|,120|,'L4'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"feature": {
"label": "feature to show",
"suggestions": "",
"help": "Feature to show in connectivity matrix; the only features applicable to groupBy='cell' are 'weight', 'delay' and 'numConns'; 'strength' = weight * probability ('weight'|'delay'|'numConns'|'probability'|'strength'|'convergence'|'divergence')g.",
"hintText": "",
"options": [
"weight",
"delay",
"numConns",
"probability",
"strength",
"convergency",
"divergency"
],
"type": "str"
},
"groupBy": {
"label": "group by",
"suggestions": "",
"help": "Show matrix for individual cells or populations ('pop'|'cell').",
"hintText": "",
"options": [
"pop",
"cell"
],
"type": "str"
},
"orderBy": {
"label": "order by",
"suggestions": "",
"help": "Unique numeric cell property to order x and y axes by, e.g. 'gid', 'ynorm', 'y' (requires groupBy='cells') ('gid'|'y'|'ynorm'|...).",
"hintText": "",
"options": [
"gid",
"y",
"ynorm"
],
"type": "str"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"granger": {
"label": "Granger",
"suggestions": "",
"help": "Calculate and optionally plot Granger Causality.",
"hintText": "",
"children": {
"cells1": {
"label": "population (or cell by index) to subset 1",
"suggestions": "",
"help": "Subset of cells from which to obtain spike train 1 (['all',|'allCells','allNetStims',|,120,|,'E1'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"cells2": {
"label": "population (or cell by index cell) to subset 2",
"suggestions": "",
"help": "Subset of cells from which to obtain spike train 2 (['all',|'allCells','allNetStims',|,120,|,'E1'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"spks1": {
"label": "spike times to train 1",
"suggestions": "",
"help": "Spike train 1; list of spike times; if omitted then obtains spikes from cells1 (list).",
"hintText": "",
"type": "list"
},
"spks2": {
"label": "spike times to train 2",
"suggestions": "",
"help": "Spike train 2; list of spike times; if omitted then obtains spikes from cells1 (list).",
"hintText": "",
"type": "list"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Range of time to calculate nTE in ms ([min, max]).",
"hintText": "",
"type": "list(float)"
},
"binSize": {
"label": "bin size",
"suggestions": "",
"help": "Bin size used to convert spike times into histogram (int).",
"hintText": "",
"type": "float"
},
"label1": {
"label": "label for train 1",
"suggestions": "",
"help": "Label for spike train 1 to use in plot (string).",
"hintText": "",
"type": "str"
},
"label2": {
"label": "label for train 2",
"suggestions": "",
"help": "Label for spike train 2 to use in plot (string).",
"hintText": "",
"type": "str"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
},
"nTE": {
"label": "Normalize Transfer Entropy",
"suggestions": "",
"help": "Calculate normalized transfer entropy.",
"hintText": "",
"children": {
"cell1": {
"label": "Cell Subset 1",
"suggestions": "",
"help": "Subset of cells from which to obtain spike train 1 (['all',|'allCells','allNetStims',|,120,|,'E1'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"cell2": {
"label": "Cell Subset 2",
"suggestions": "",
"help": "Subset of cells from which to obtain spike train 2 (['all',|'allCells','allNetStims',|,120,|,'E1'|,('L2', 56)|,('L5',[4,5,6])]).",
"hintText": "",
"type": "list"
},
"spks1": {
"label": "Spike train 1",
"suggestions": "",
"help": "Spike train 1; list of spike times; if omitted then obtains spikes from cells1 (list).",
"hintText": "",
"type": "list(float)"
},
"spks2": {
"label": "Spike train 2",
"suggestions": "",
"help": "Spike train 2; list of spike times; if omitted then obtains spikes from cells1 (list).",
"hintText": "",
"type": "list(float)"
},
"timeRange": {
"label": "Time range [min,max] (ms)",
"suggestions": "",
"help": "Range of time to calculate nTE in ms ([min, max]).",
"hintText": "",
"type": "list(float)"
},
"binSize": {
"label": "Bin size",
"suggestions": "",
"help": "Bin size used to convert spike times into histogram (int).",
"hintText": "",
"type": "float"
},
"numShuffle": {
"label": "Number of Shuffles",
"suggestions": "",
"help": "Number of times to shuffle spike train 1 to calculate TEshuffled; note: nTE = (TE - TEShuffled)/H(X2F|X2P) (int).",
"hintText": "",
"type": "float"
},
"figSize": {
"label": "Figure size",
"suggestions": "",
"help": "Size of figure ((width, height))",
"hintText": "",
"type": ""
},
"saveData": {
"label": "Save data",
"suggestions": "",
"help": "File name where to save the final data used to generate the figure (None|'fileName').",
"hintText": "",
"type": "str"
},
"saveFig": {
"label": "Save figure file name",
"suggestions": "",
"help": "File name where to save the figure (None|'fileName')",
"hintText": "",
"type": "str"
},
"showFig": {
"label": "Show figure",
"suggestions": "",
"help": "Whether to show the figure or not (True|False).",
"hintText": "",
"type": "bool"
}
}
}
}
}
}
}
}
| 53.659082 | 467 | 0.323153 | [
"MIT"
] | Anjali-Agarwal8/netpyne | netpyne/metadata/metadata.py | 132,055 | Python |
#!/usr/bin/env python3
# Copyright (c) 2014-2017 The Bitcoin Core developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Test mining RPCs
- getmininginfo
- getblocktemplate proposal mode
- submitblock"""
import copy
from binascii import b2a_hex
from decimal import Decimal
from test_framework.blocktools import create_coinbase
from test_framework.mininode import CBlock
from test_framework.test_framework import LivecoinTestFramework
from test_framework.util import assert_equal, assert_raises_rpc_error
def b2x(b):
return b2a_hex(b).decode('ascii')
def assert_template(node, block, expect, rehash=True):
if rehash:
block.hashMerkleRoot = block.calc_merkle_root()
rsp = node.getblocktemplate({'data': b2x(block.serialize()), 'mode': 'proposal'})
assert_equal(rsp, expect)
class MiningTest(LivecoinTestFramework):
def set_test_params(self):
self.num_nodes = 2
self.setup_clean_chain = False
def run_test(self):
node = self.nodes[0]
self.log.info('getmininginfo')
mining_info = node.getmininginfo()
assert_equal(mining_info['blocks'], 200)
assert_equal(mining_info['chain'], 'regtest')
assert_equal(mining_info['currentblocktx'], 0)
assert_equal(mining_info['currentblockweight'], 0)
assert_equal(mining_info['difficulty'], Decimal('4.656542373906925E-10'))
assert_equal(mining_info['networkhashps'], Decimal('0.003333333333333334'))
assert_equal(mining_info['pooledtx'], 0)
# Mine a block to leave initial block download
node.generate(1)
tmpl = node.getblocktemplate()
self.log.info("getblocktemplate: Test capability advertised")
assert 'proposal' in tmpl['capabilities']
assert 'coinbasetxn' not in tmpl
coinbase_tx = create_coinbase(height=int(tmpl["height"]) + 1)
# sequence numbers must not be max for nLockTime to have effect
coinbase_tx.vin[0].nSequence = 2 ** 32 - 2
coinbase_tx.rehash()
block = CBlock()
block.nVersion = tmpl["version"]
block.hashPrevBlock = int(tmpl["previousblockhash"], 16)
block.nTime = tmpl["curtime"]
block.nBits = int(tmpl["bits"], 16)
block.nNonce = 0
block.vtx = [coinbase_tx]
self.log.info("getblocktemplate: Test valid block")
assert_template(node, block, None)
self.log.info("submitblock: Test block decode failure")
assert_raises_rpc_error(-22, "Block decode failed", node.submitblock, b2x(block.serialize()[:-15]))
self.log.info("getblocktemplate: Test bad input hash for coinbase transaction")
bad_block = copy.deepcopy(block)
bad_block.vtx[0].vin[0].prevout.hash += 1
bad_block.vtx[0].rehash()
assert_template(node, bad_block, 'bad-cb-missing')
self.log.info("submitblock: Test invalid coinbase transaction")
assert_raises_rpc_error(-22, "Block does not start with a coinbase", node.submitblock, b2x(bad_block.serialize()))
self.log.info("getblocktemplate: Test truncated final transaction")
assert_raises_rpc_error(-22, "Block decode failed", node.getblocktemplate, {'data': b2x(block.serialize()[:-1]), 'mode': 'proposal'})
self.log.info("getblocktemplate: Test duplicate transaction")
bad_block = copy.deepcopy(block)
bad_block.vtx.append(bad_block.vtx[0])
assert_template(node, bad_block, 'bad-txns-duplicate')
self.log.info("getblocktemplate: Test invalid transaction")
bad_block = copy.deepcopy(block)
bad_tx = copy.deepcopy(bad_block.vtx[0])
bad_tx.vin[0].prevout.hash = 255
bad_tx.rehash()
bad_block.vtx.append(bad_tx)
assert_template(node, bad_block, 'bad-txns-inputs-missingorspent')
self.log.info("getblocktemplate: Test nonfinal transaction")
bad_block = copy.deepcopy(block)
bad_block.vtx[0].nLockTime = 2 ** 32 - 1
bad_block.vtx[0].rehash()
assert_template(node, bad_block, 'bad-txns-nonfinal')
self.log.info("getblocktemplate: Test bad tx count")
# The tx count is immediately after the block header
TX_COUNT_OFFSET = 80
bad_block_sn = bytearray(block.serialize())
assert_equal(bad_block_sn[TX_COUNT_OFFSET], 1)
bad_block_sn[TX_COUNT_OFFSET] += 1
assert_raises_rpc_error(-22, "Block decode failed", node.getblocktemplate, {'data': b2x(bad_block_sn), 'mode': 'proposal'})
self.log.info("getblocktemplate: Test bad bits")
bad_block = copy.deepcopy(block)
bad_block.nBits = 469762303 # impossible in the real world
assert_template(node, bad_block, 'bad-diffbits')
self.log.info("getblocktemplate: Test bad merkle root")
bad_block = copy.deepcopy(block)
bad_block.hashMerkleRoot += 1
assert_template(node, bad_block, 'bad-txnmrklroot', False)
self.log.info("getblocktemplate: Test bad timestamps")
bad_block = copy.deepcopy(block)
bad_block.nTime = 2 ** 31 - 1
assert_template(node, bad_block, 'time-too-new')
bad_block.nTime = 0
assert_template(node, bad_block, 'time-too-old')
self.log.info("getblocktemplate: Test not best block")
bad_block = copy.deepcopy(block)
bad_block.hashPrevBlock = 123
assert_template(node, bad_block, 'inconclusive-not-best-prevblk')
if __name__ == '__main__':
MiningTest().main()
| 41.058824 | 141 | 0.681411 | [
"MIT"
] | elliottminns/livecoin | test/functional/mining_basic.py | 5,584 | Python |
import tarfile
import textwrap
import pytoml
from .app import App
from .exceptions import TockLoaderException
from .tbfh import TBFHeader
class TAB:
'''
Tock Application Bundle object. This class handles the TAB format.
'''
def __init__ (self, tab_path):
self.tab = tarfile.open(tab_path)
def extract_app (self, arch):
'''
Return an `App` object from this TAB. You must specify the desired
MCU architecture so the correct binary can be retrieved.
'''
binary_tarinfo = self.tab.getmember('{}.bin'.format(arch))
binary = self.tab.extractfile(binary_tarinfo).read()
# First get the TBF header from the correct binary in the TAB
tbfh = TBFHeader(binary)
if tbfh.is_valid():
name_or_params = tbfh.get_app_name()
if isinstance(name_or_params, str):
name = name_or_params
else:
start = name_or_params[0]
end = start+name_or_params[1]
name = binary[start:end].decode('utf-8')
# Check that total size actually matches the binary that we got.
if tbfh.get_app_size() < len(binary):
# It's fine if the binary is smaller, but the binary cannot be
# longer than the amount of reserved space (`total_size` in the
# TBF header) for the app.
raise TockLoaderException('Invalid TAB, the app binary is longer than its defined total_size')
return App(tbfh, None, name, binary)
else:
raise TockLoaderException('Invalid TBF found in app in TAB')
def is_compatible_with_board (self, board):
'''
Check if the Tock app is compatible with a particular Tock board.
'''
metadata = self.parse_metadata()
if metadata['tab-version'] == 1:
return 'only-for-boards' not in metadata or \
board in metadata['only-for-boards'] or \
metadata['only-for-boards'] == ''
else:
raise TockLoaderException('Unable to understand version {} of metadata'.format(metadata['tab-version']))
def parse_metadata (self):
'''
Open and parse the included metadata file in the TAB.
'''
metadata_tarinfo = self.tab.getmember('metadata.toml')
metadata_str = self.tab.extractfile(metadata_tarinfo).read().decode('utf-8')
return pytoml.loads(metadata_str)
def get_supported_architectures (self):
'''
Return a list of architectures that this TAB has compiled binaries for.
'''
contained_files = self.tab.getnames()
return [i[:-4] for i in contained_files if i[-4:] == '.bin']
def get_tbf_header (self):
'''
Return a TBFHeader object with the TBF header from the app in the TAB.
TBF headers are not architecture specific, so we pull from a random
binary if there are multiple architectures supported.
'''
# Find a .bin file
for f in self.tab.getnames():
if f[-4:] == '.bin':
binary_tarinfo = self.tab.getmember(f)
binary = self.tab.extractfile(binary_tarinfo).read()
# Get the TBF header from a binary in the TAB
return TBFHeader(binary)
return None
def __str__ (self):
out = ''
metadata = self.parse_metadata()
out += 'TAB: {}\n'.format(metadata['name'])
for k,v in sorted(metadata.items()):
if k == 'name':
continue
out += ' {}: {}\n'.format(k,v)
out += ' supported architectures: {}\n'.format(', '.join(self.get_supported_architectures()))
out += ' TBF Header\n'
out += textwrap.indent(str(self.get_tbf_header()), ' ')
return out
| 32.009709 | 107 | 0.693358 | [
"MIT"
] | torfmaster/tockloader | tockloader/tab.py | 3,297 | Python |
from __future__ import absolute_import
from __future__ import unicode_literals
import types
import copy
from django import forms
from django.forms.forms import NON_FIELD_ERRORS
from django.core.validators import EMPTY_VALUES
from django.db import models
from django.db.models.fields import FieldDoesNotExist
from django.utils import six
from django.utils.text import capfirst
from django.utils.translation import ugettext as _
from sys import version_info
try:
from django.db.models.constants import LOOKUP_SEP
except ImportError: # pragma: nocover
# Django < 1.5 fallback
from django.db.models.sql.constants import LOOKUP_SEP # noqa
try:
from collections import OrderedDict
except ImportError: # pragma: nocover
# Django < 1.5 fallback
from django.utils.datastructures import SortedDict as OrderedDict # noqa
try:
from django.db.models.related import RelatedObject as ForeignObjectRel
except ImportError: # pragma: nocover
# Django >= 1.8 replaces RelatedObject with ForeignObjectRel
from django.db.models.fields.related import ForeignObjectRel
from .filters import (Filter, CharFilter, BooleanFilter,
ChoiceFilter, DateFilter, DateTimeFilter, TimeFilter, ModelChoiceFilter,
ModelMultipleChoiceFilter, NumberFilter)
ORDER_BY_FIELD = 'o'
# There is a bug with deepcopy in 2.6, patch if we are running python < 2.7
# http://bugs.python.org/issue1515
if version_info < (2, 7, 0):
def _deepcopy_method(x, memo):
return type(x)(x.im_func, copy.deepcopy(x.im_self, memo), x.im_class)
copy._deepcopy_dispatch[types.MethodType] = _deepcopy_method
class STRICTNESS(object):
"""
Values of False & True chosen for backward compatability reasons.
Originally, these were the only options.
"""
IGNORE = False
RETURN_NO_RESULTS = True
RAISE_VALIDATION_ERROR = "RAISE"
def get_declared_filters(bases, attrs, with_base_filters=True):
filters = []
for filter_name, obj in list(attrs.items()):
if isinstance(obj, Filter):
obj = attrs.pop(filter_name)
if getattr(obj, 'name', None) is None:
obj.name = filter_name
filters.append((filter_name, obj))
filters.sort(key=lambda x: x[1].creation_counter)
if with_base_filters:
for base in bases[::-1]:
if hasattr(base, 'base_filters'):
filters = list(base.base_filters.items()) + filters
else:
for base in bases[::-1]:
if hasattr(base, 'declared_filters'):
filters = list(base.declared_filters.items()) + filters
return OrderedDict(filters)
def get_model_field(model, f):
parts = f.split(LOOKUP_SEP)
opts = model._meta
for name in parts[:-1]:
try:
rel = opts.get_field_by_name(name)[0]
except FieldDoesNotExist:
return None
if isinstance(rel, ForeignObjectRel):
if hasattr(rel, "related_model"):
# django >= 1.8 (ForeignObjectRel)
opts = rel.related_model._meta
else:
# django < 1.8 (RelatedObject)
opts = rel.opts
else:
model = rel.rel.to
opts = model._meta
try:
rel, model, direct, m2m = opts.get_field_by_name(parts[-1])
except FieldDoesNotExist:
return None
return rel
def filters_for_model(model, fields=None, exclude=None, filter_for_field=None,
filter_for_reverse_field=None):
field_dict = OrderedDict()
opts = model._meta
if fields is None:
fields = [f.name for f in sorted(opts.fields + opts.many_to_many)
if not isinstance(f, models.AutoField)]
# Loop through the list of fields.
for f in fields:
# Skip the field if excluded.
if exclude is not None and f in exclude:
continue
field = get_model_field(model, f)
# Do nothing if the field doesn't exist.
if field is None:
field_dict[f] = None
continue
if isinstance(field, ForeignObjectRel):
filter_ = filter_for_reverse_field(field, f)
if filter_:
field_dict[f] = filter_
# If fields is a dictionary, it must contain lists.
elif isinstance(fields, dict):
# Create a filter for each lookup type.
for lookup_type in fields[f]:
filter_ = filter_for_field(field, f, lookup_type)
if filter_:
filter_name = f
# Don't add "exact" to filter names
if lookup_type != 'exact':
filter_name = f + LOOKUP_SEP + lookup_type
field_dict[filter_name] = filter_
# If fields is a list, it contains strings.
else:
filter_ = filter_for_field(field, f)
if filter_:
field_dict[f] = filter_
return field_dict
def get_full_clean_override(together):
def full_clean(form):
def add_error(message):
try:
form.add_error(None, message)
except AttributeError:
form._errors[NON_FIELD_ERRORS] = message
def all_valid(fieldset):
cleaned_data = form.cleaned_data
count = len([i for i in fieldset if cleaned_data.get(i)])
return 0 < count < len(fieldset)
super(form.__class__, form).full_clean()
message = 'Following fields must be together: %s'
if isinstance(together[0], (list, tuple)):
for each in together:
if all_valid(each):
return add_error(message % ','.join(each))
elif all_valid(together):
return add_error(message % ','.join(together))
return full_clean
class FilterSetOptions(object):
def __init__(self, options=None):
self.model = getattr(options, 'model', None)
self.fields = getattr(options, 'fields', None)
self.exclude = getattr(options, 'exclude', None)
self.order_by = getattr(options, 'order_by', False)
self.form = getattr(options, 'form', forms.Form)
self.together = getattr(options, 'together', None)
class FilterSetMetaclass(type):
def __new__(cls, name, bases, attrs):
try:
parents = [b for b in bases if issubclass(b, FilterSet)]
except NameError:
# We are defining FilterSet itself here
parents = None
declared_filters = get_declared_filters(bases, attrs, False)
new_class = super(
FilterSetMetaclass, cls).__new__(cls, name, bases, attrs)
if not parents:
return new_class
opts = new_class._meta = FilterSetOptions(
getattr(new_class, 'Meta', None))
if opts.model:
filters = filters_for_model(opts.model, opts.fields, opts.exclude,
new_class.filter_for_field,
new_class.filter_for_reverse_field)
filters.update(declared_filters)
else:
filters = declared_filters
if None in filters.values():
raise TypeError("Meta.fields contains a field that isn't defined "
"on this FilterSet")
new_class.declared_filters = declared_filters
new_class.base_filters = filters
return new_class
FILTER_FOR_DBFIELD_DEFAULTS = {
models.AutoField: {
'filter_class': NumberFilter
},
models.CharField: {
'filter_class': CharFilter
},
models.TextField: {
'filter_class': CharFilter
},
models.BooleanField: {
'filter_class': BooleanFilter
},
models.DateField: {
'filter_class': DateFilter
},
models.DateTimeField: {
'filter_class': DateTimeFilter
},
models.TimeField: {
'filter_class': TimeFilter
},
models.OneToOneField: {
'filter_class': ModelChoiceFilter,
'extra': lambda f: {
'queryset': f.rel.to._default_manager.complex_filter(
f.rel.limit_choices_to),
'to_field_name': f.rel.field_name,
}
},
models.ForeignKey: {
'filter_class': ModelChoiceFilter,
'extra': lambda f: {
'queryset': f.rel.to._default_manager.complex_filter(
f.rel.limit_choices_to),
'to_field_name': f.rel.field_name
}
},
models.ManyToManyField: {
'filter_class': ModelMultipleChoiceFilter,
'extra': lambda f: {
'queryset': f.rel.to._default_manager.complex_filter(
f.rel.limit_choices_to),
}
},
models.DecimalField: {
'filter_class': NumberFilter,
},
models.SmallIntegerField: {
'filter_class': NumberFilter,
},
models.IntegerField: {
'filter_class': NumberFilter,
},
models.PositiveIntegerField: {
'filter_class': NumberFilter,
},
models.PositiveSmallIntegerField: {
'filter_class': NumberFilter,
},
models.FloatField: {
'filter_class': NumberFilter,
},
models.NullBooleanField: {
'filter_class': BooleanFilter,
},
models.SlugField: {
'filter_class': CharFilter,
},
models.EmailField: {
'filter_class': CharFilter,
},
models.FilePathField: {
'filter_class': CharFilter,
},
models.URLField: {
'filter_class': CharFilter,
},
models.IPAddressField: {
'filter_class': CharFilter,
},
models.CommaSeparatedIntegerField: {
'filter_class': CharFilter,
},
}
class BaseFilterSet(object):
filter_overrides = {}
order_by_field = ORDER_BY_FIELD
# What to do on on validation errors
strict = STRICTNESS.RETURN_NO_RESULTS
def __init__(self, data=None, queryset=None, prefix=None, strict=None):
self.is_bound = data is not None
self.data = data or {}
if queryset is None:
queryset = self._meta.model._default_manager.all()
self.queryset = queryset
self.form_prefix = prefix
if strict is not None:
self.strict = strict
self.filters = copy.deepcopy(self.base_filters)
# propagate the model being used through the filters
for filter_ in self.filters.values():
filter_.model = self._meta.model
# Apply the parent to the filters, this will allow the filters to access the filterset
for filter_key, filter_ in six.iteritems(self.filters):
filter_.parent = self
def __iter__(self):
for obj in self.qs:
yield obj
def __len__(self):
return len(self.qs)
def __getitem__(self, key):
return self.qs[key]
@property
def qs(self):
if not hasattr(self, '_qs'):
valid = self.is_bound and self.form.is_valid()
if self.is_bound and not valid:
if self.strict == STRICTNESS.RAISE_VALIDATION_ERROR:
raise forms.ValidationError(self.form.errors)
elif bool(self.strict) == STRICTNESS.RETURN_NO_RESULTS:
self._qs = self.queryset.none()
return self._qs
# else STRICTNESS.IGNORE... ignoring
# start with all the results and filter from there
qs = self.queryset.all()
for name, filter_ in six.iteritems(self.filters):
value = None
if valid:
value = self.form.cleaned_data[name]
else:
raw_value = self.form[name].value()
try:
value = self.form.fields[name].clean(raw_value)
except forms.ValidationError:
if self.strict == STRICTNESS.RAISE_VALIDATION_ERROR:
raise
elif bool(self.strict) == STRICTNESS.RETURN_NO_RESULTS:
self._qs = self.queryset.none()
return self._qs
# else STRICTNESS.IGNORE... ignoring
if value is not None: # valid & clean data
qs = filter_.filter(qs, value)
if self._meta.order_by:
order_field = self.form.fields[self.order_by_field]
data = self.form[self.order_by_field].data
ordered_value = None
try:
ordered_value = order_field.clean(data)
except forms.ValidationError:
pass
if ordered_value in EMPTY_VALUES and self.strict:
ordered_value = self.form.fields[self.order_by_field].choices[0][0]
if ordered_value:
qs = qs.order_by(*self.get_order_by(ordered_value))
self._qs = qs
return self._qs
def count(self):
return self.qs.count()
@property
def form(self):
if not hasattr(self, '_form'):
fields = OrderedDict([
(name, filter_.field)
for name, filter_ in six.iteritems(self.filters)])
fields[self.order_by_field] = self.ordering_field
Form = type(str('%sForm' % self.__class__.__name__),
(self._meta.form,), fields)
if self._meta.together:
Form.full_clean = get_full_clean_override(self._meta.together)
if self.is_bound:
self._form = Form(self.data, prefix=self.form_prefix)
else:
self._form = Form(prefix=self.form_prefix)
return self._form
def get_ordering_field(self):
if self._meta.order_by:
if isinstance(self._meta.order_by, (list, tuple)):
if isinstance(self._meta.order_by[0], (list, tuple)):
# e.g. (('field', 'Display name'), ...)
choices = [(f[0], f[1]) for f in self._meta.order_by]
else:
choices = [(f, _('%s (descending)' % capfirst(f[1:])) if f[0] == '-' else capfirst(f))
for f in self._meta.order_by]
else:
# add asc and desc field names
# use the filter's label if provided
choices = []
for f, fltr in self.filters.items():
choices.extend([
(fltr.name or f, fltr.label or capfirst(f)),
("-%s" % (fltr.name or f), _('%s (descending)' % (fltr.label or capfirst(f))))
])
return forms.ChoiceField(label=_("Ordering"), required=False,
choices=choices)
@property
def ordering_field(self):
if not hasattr(self, '_ordering_field'):
self._ordering_field = self.get_ordering_field()
return self._ordering_field
def get_order_by(self, order_choice):
return [order_choice]
@classmethod
def filter_for_field(cls, f, name, lookup_type='exact'):
filter_for_field = dict(FILTER_FOR_DBFIELD_DEFAULTS)
filter_for_field.update(cls.filter_overrides)
default = {
'name': name,
'label': capfirst(f.verbose_name),
'lookup_type': lookup_type
}
if f.choices:
default['choices'] = f.choices
return ChoiceFilter(**default)
data = filter_for_field.get(f.__class__)
if data is None:
# could be a derived field, inspect parents
for class_ in f.__class__.mro():
# skip if class_ is models.Field or object
# 1st item in mro() is original class
if class_ in (f.__class__, models.Field, object):
continue
data = filter_for_field.get(class_)
if data:
break
if data is None:
return
filter_class = data.get('filter_class')
default.update(data.get('extra', lambda f: {})(f))
if filter_class is not None:
return filter_class(**default)
@classmethod
def filter_for_reverse_field(cls, f, name):
rel = f.field.rel
queryset = f.field.model._default_manager.all()
default = {
'name': name,
'label': capfirst(rel.related_name),
'queryset': queryset,
}
if rel.multiple:
return ModelMultipleChoiceFilter(**default)
else:
return ModelChoiceFilter(**default)
class FilterSet(six.with_metaclass(FilterSetMetaclass, BaseFilterSet)):
pass
def filterset_factory(model):
meta = type(str('Meta'), (object,), {'model': model})
filterset = type(str('%sFilterSet' % model._meta.object_name),
(FilterSet,), {'Meta': meta})
return filterset
| 33.884692 | 106 | 0.582962 | [
"BSD-3-Clause"
] | aioTV/django-filter | django_filters/filterset.py | 17,044 | Python |
from BiblioAlly import catalog as cat, domain, translator as bibtex
class IeeeXTranslator(bibtex.Translator):
def _document_from_proto_document(self, proto_document):
bibtex.Translator._translate_kind(proto_document)
kind = proto_document['type']
fields = proto_document['field']
if 'title' in fields:
title = self._unbroken(self._uncurlied(fields['title']))
else:
title = ''
if 'abstract' in fields:
abstract = self._unbroken(self._uncurlied(fields['abstract']))
else:
abstract = ''
year = int(fields['year'])
author_field = ''
if 'author' in fields:
author_field = self._unbroken(self._all_uncurly(fields['author'].replace('}and', ' and')))
if author_field == '':
author_field = 'Author, Unamed'
authors = self._authors_from_field(author_field)
affiliations = self._expand_affiliations(None, authors)
keywords = []
if 'keywords' in fields:
all_keywords = self._all_uncurly(fields['keywords']).split(';')
keyword_names = set()
for keyword_name in all_keywords:
sub_keyword_names = keyword_name.split(',')
for sub_keyword_name in sub_keyword_names:
name = sub_keyword_name.strip().capitalize()
if name not in keyword_names:
keyword_names.add(name)
keyword_names = list(keyword_names)
for keyword_name in keyword_names:
keywords.append(domain.Keyword(name=keyword_name))
document = domain.Document(proto_document['id'].strip(), kind, title, abstract, keywords, year, affiliations)
document.generator = "IEEE Xplore"
if 'doi' in fields:
document.doi = self._uncurlied(fields['doi'])
if 'journal' in fields:
document.journal = self._uncurlied(fields['journal'])
elif 'booktitle' in fields and kind == 'inproceedings':
document.journal = self._uncurlied(fields['booktitle'])
if 'number' in fields:
if len(self._uncurlied(fields['number'])) > 0:
document.number = self._uncurlied(fields['number'])
if 'pages' in fields:
if len(self._uncurlied(fields['pages'])) > 0:
document.pages = self._uncurlied(fields['pages'])
if 'url' in fields:
if len(self._uncurlied(fields['url'])) > 0:
document.url = self._uncurlied(fields['url'])
if 'volume' in fields:
if len(self._uncurlied(fields['volume'])) > 0:
document.volume = self._uncurlied(fields['volume'])
return document
def _proto_document_from_document(self, document: domain.Document):
kind = document.kind
if kind == 'proceedings':
kind = 'inproceedings'
fields = dict()
fields['external_key'] = document.external_key
doc_authors = document.authors
doc_authors.sort(key=lambda doc_author: doc_author.first)
doc_authors.reverse()
all_authors = [(doc_author.author.long_name if doc_author.author.long_name is not None
else doc_author.author.short_name) for doc_author in doc_authors]
fields['author'] = self._curly(all_authors, separator=' and ')
if document.journal is not None:
if document.kind == 'article':
fields['journal'] = self._curly(str(document.journal))
else:
fields['booktitle'] = self._curly(str(document.journal))
fields['title'] = self._curly(document.title)
affiliations = []
for doc_author in doc_authors:
institution = doc_author.institution
if institution is not None:
affiliation = ', '.join([institution.name, institution.country])
affiliations.append(affiliation)
if len(affiliations) > 0:
fields['affiliation'] = self._curly(affiliations, '; ')
fields['year'] = self._curly(str(document.year))
if document.international_number is not None:
fields['ISSN'] = self._curly(str(document.international_number))
if document.publisher is not None:
fields['publisher'] = self._curly(str(document.publisher))
if document.address is not None:
fields['address'] = self._curly(str(document.address))
if document.doi is not None:
fields['doi'] = self._curly(str(document.doi))
if document.international_number is not None:
fields['url'] = self._curly(str(document.url))
fields['abstract'] = self._curly(document.abstract)
if document.pages is not None:
fields['pages'] = self._curly(str(document.pages))
if document.volume is not None:
fields['volume'] = self._curly(str(document.volume))
if document.number is not None:
fields['number'] = self._curly(str(document.number))
if document.language is not None:
fields['language'] = self._curly(str(document.language))
keywords = [keyword.name for keyword in document.keywords]
fields['keywords'] = self._curly(keywords, ';')
if len(document.references) > 0:
fields['references'] = self._curly('; '.join(document.references))
if document.document_type is not None:
fields['document_type'] = self._curly(document.document_type)
fields['source'] = self._curly(document.generator)
proto_document = {
'type': kind,
'fields': fields
}
return proto_document
def _as_bibtex(self, proto_document):
kind = proto_document['type'].upper()
fields = proto_document['fields']
external_key = fields['external_key']
del fields['external_key']
key_value = []
for key, value in fields.items():
key_value.append(f'{key}={value}')
bibtex = f'@{kind}' + '{' + f'{external_key},\n' + ',\n'.join(key_value) + '\n}'
return bibtex
IeeeXplore = "IeeeXplore"
cat.Catalog.translators[IeeeXplore] = IeeeXTranslator
| 44.992806 | 117 | 0.605373 | [
"MIT"
] | gambit4348/BiblioAlly | BiblioAlly/ieee.py | 6,254 | Python |
"""empty message
Revision ID: 780c29109b25
Revises: 911cc5d772fc
Create Date: 2020-08-30 15:22:15.026266
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = "780c29109b25"
down_revision = "911cc5d772fc"
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_foreign_key(None, "feedback", "user", ["author_id"], ["id"])
op.drop_column("feedback", "author")
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column("feedback", sa.Column("author", sa.VARCHAR(length=20), nullable=True))
op.drop_constraint(None, "feedback", type_="foreignkey")
# ### end Alembic commands ###
| 26.806452 | 89 | 0.665463 | [
"MIT"
] | mutalisk999/Flog | migrations/versions/780c29109b25_.py | 831 | Python |
import numpy as np
import os
import cv2
from PIL import Image
import numpy as np
import random
import itertools
import matplotlib.pyplot as plt # plt 用于显示图片
from tqdm import tqdm
# 标注文件数据处理
def read_pslot(annt_file):
# print(annt_file)
with open(annt_file, "r") as f:
annt = f.readlines()
# print("annt", annt)
l = []
l_ign = []
for line in annt:
line_annt = line.strip('\n').split(' ')
# print(line_annt)
if len(line_annt) != 13 or line_annt[0] != 'line' or line_annt[-4] == '3':
continue
if line_annt[-4] in ['0', '1']:
l.append(np.array([int(line_annt[i + 1]) for i in range(8)]))
# continue
# if line_annt[-4] in ['1', '5']:
# l_ign.append(np.array([int(line_annt[i + 1]) for i in range(8)]))
# continue
return l, l_ign
# 标点
def colorize(points_list, img, save_path, item, line, point_color):
save_path = os.path.join(save_path, str(
item.strip('.jpg'))+"_"+str(line)+".jpg")
img2 = img.copy()
# print(save_path)
# points_list = 384 * np.abs(np.array(outputs[0], dtype=np.float))
point_size = 1
thickness = 4 # 可以为 0、4、8
for i in range(4):
cv2.circle(img2, (int(points_list[i][0]), int(points_list[i][1])),
point_size, point_color, thickness)
# print(save_path)
cv2.imwrite(save_path, img2)
# 画线
def paint_line(img, dst, cropimg_path, num):
img2 = img.copy()
cv2.line(img2, (int(dst[0][0]), int(dst[0][1])), (int(
dst[1][0]), int(dst[1][1])), (255, 0, 0), 5)
cv2.line(img2, (int(dst[1][0]), int(dst[1][1])), (int(
dst[2][0]), int(dst[2][1])), (255, 0, 0), 5)
cv2.line(img2, (int(dst[2][0]), int(dst[2][1])), (int(
dst[3][0]), int(dst[3][1])), (255, 0, 0), 5)
cv2.line(img2, (int(dst[3][0]), int(dst[3][1])), (int(
dst[0][0]), int(dst[0][1])), (255, 0, 0), 5)
cropimg_path1 = os.path.join(
cropimg_path, i.strip('.jpg')+'_'+str(num)+'.jpg')
cv2.imwrite(cropimg_path1, img2)
def Crop_pic(ps, img_path, cropimg_path, perspective_path, txt_file, i, trans_path, save_path1, save_path2):
# single pic
img = cv2.imread(img_path)
perspective3 = np.float32([[0, 0], [383, 0], [383, 383], [0, 383]])
perspective3_ = np.float32([[0, 0], [383, 0], [383, 383]])
num = 0
for line in ps:
num = num + 1
# 随机生成4个坐标
arr0 = random.randint(80, 120)
arr1 = random.randint(80, 120)
arr2 = random.randint(263, 303)
arr3 = random.randint(80, 120)
arr4 = random.randint(263, 303)
arr5 = random.randint(263, 303)
arr6 = random.randint(80, 120)
arr7 = random.randint(263, 303)
perspective0 = np.float32([[line[0], line[1]], [line[2], line[3]], [
line[4], line[5]], [line[6], line[7]]])
perspective0_ = np.float32([[line[0], line[1]], [line[2], line[3]], [
line[4], line[5]]])
colorize(perspective0, img, save_path1, i, num, (0, 255, 0))
perspective1 = np.float32(
[[arr0, arr1], [arr2, arr3], [arr4, arr5], [arr6, arr7]])
perspective1_ = np.float32(
[[arr0, arr1], [arr2, arr3], [arr4, arr5]])
# 求逆变换矩阵
# trans_inv = cv2.getPerspectiveTransform(perspective1, perspective0)
trans_inv = cv2.getAffineTransform(perspective1_, perspective0_)
# 求逆投影变换后的点坐标
dst = []
# mat = np.array(
# [[[0, 0], [383, 0], [383, 383], [0, 383]]], dtype=np.float32)
mat = np.array(
[[0, 0, 1], [383, 0, 1], [383, 383, 1], [0, 383, 1]], dtype=np.float32)
mat = mat.transpose()
# dst = cv2.perspectiveTransform(mat, trans_inv)
dst = np.dot(trans_inv, mat)
dst = dst.transpose()
# 画线
paint_line(img, dst, cropimg_path, num)
# 将停车位投影变换后得到在384*384分辨率下的停车位图像
# perspective2 = np.float32([[dst[0][0][0], dst[0][0][1]], [dst[0][1][0], dst[0][1][1]], [
# dst[0][2][0], dst[0][2][1]], [dst[0][3][0], dst[0][3][1]]])
perspective2_ = np.float32([[dst[0][0], dst[0][1]], [dst[1][0], dst[1][1]], [
dst[2][0], dst[2][1]]])
# trans = cv2.getPerspectiveTransform(perspective2, perspective3)
# dst2 = cv2.warpPerspective(img, trans, (384, 384))
trans = cv2.getAffineTransform(perspective2_, perspective3_)
dst2 = cv2.warpAffine(img, trans, (384, 384))
# 保存原图四个内角点在384*384图片上的坐标
# mat2 = np.array([[[line[0], line[1]], [line[2], line[3]], [
# line[4], line[5]], [line[6], line[7]]]], dtype=np.float32)
mat2 = np.array([[line[0], line[1], 1], [line[2], line[3], 1], [
line[4], line[5], 1], [line[6], line[7], 1]], dtype=np.float32)
mat2 = mat2.transpose()
point = np.dot(trans, mat2)
point = point.transpose()
# point = cv2.perspectiveTransform(mat2, trans)
# point = np.dot(mat2, trans)
perspective_path1 = os.path.join(
perspective_path, i.strip('.jpg')+'_'+str(num)+'.jpg')
# print(perspective_path)
cv2.imwrite(perspective_path1, dst2)
colorize(point, dst2, save_path2, i, num, (0, 255, 0))
# 把四个坐标点记录下来
txt_file1 = os.path.join(
txt_file, i.strip('.jpg')+'_'+str(num)+'_OA.txt')
with open(txt_file1, "w") as f:
for j in range(4):
f.write(str(point[j][0]))
f.write(' ')
f.write(str(point[j][1]))
f.write('\n')
# 把转换矩阵记录下来
trans_path1 = os.path.join(
trans_path, i.strip('.jpg')+'_'+str(num)+'.txt')
with open(trans_path1, "w") as ff:
for j in range(2):
for k in range(3):
ff.write(str(trans_inv[j][k]))
ff.write(" ")
# 计算四个点的预测点与真值点之间的误差
def get_acc(y, y_hat, dis):
total = 0
total = 0
for i in range(4):
total += ((y[i][0]-y_hat[i][0])**2 + (y[i][1]-y_hat[i][1])**2)**0.5
total /= 4
if total < dis:
return 1
else:
return 0
def output_pic(img_path, output_path, trans_path, fina_path, ps2, pix, point_path):
img_pred = cv2.imread(img_path)
point_pred = []
trans_inv = []
point_pred = np.loadtxt(output_path)
point_pred = 384*np.expand_dims(point_pred, axis=0)
trans_inv = np.loadtxt(trans_path)
trans_inv = trans_inv.reshape(3, 3)
trans_inv = np.mat(trans_inv)
point_ground = np.loadtxt(point_path)
point_ground = np.expand_dims(point_ground, axis=0)
point_ground2 = cv2.perspectiveTransform(point_ground, trans_inv)
point_size = 1
thickness = 4
for i in range(4):
cv2.circle(img_pred, (int(point_ground2[0][i][0]), int(point_ground2[0][i][1])),
point_size, (0, 255, 0), thickness)
cv2.imwrite(fina_path, img_pred)
point_pred2 = cv2.perspectiveTransform(point_pred, trans_inv)
# 红色
point_color = (0, 0, 255)
point_color2 = (0, 255, 0)
for i in range(4):
cv2.circle(img_pred, (int(point_pred2[0][i][0]), int(point_pred2[0][i][1])),
point_size, point_color, thickness)
cv2.imwrite(fina_path, img_pred)
point_pred3 = point_pred2[0]
ps2 = ps2[0].reshape(4, 2)
tmp = get_acc(point_pred3, point_ground2[0], pix)
return tmp
# 精度
def output(pix):
accuracy = 0
for i in os.listdir(test_dir):
output_path = os.path.join(
"/media/alpha4TB/ziqi/Parking/CNN/output", i.strip('.jpg')+'.txt')
img_path = os.path.join(
"/media/alpha4TB/ziqi/Parking/Ps_locate_dataset/img", i)
trans_inv = os.path.join(
"/media/alpha4TB/ziqi/Parking/Ps_locate_dataset/trans_inv", i.strip('.jpg')+'.txt')
fina_path = os.path.join(
"/media/alpha4TB/ziqi/Parking/Ps_locate_dataset/fina", i)
annt_path2 = os.path.join(
'./Ps_locate_dataset/annt', i.strip('.jpg')+'_OA.txt')
point_path = os.path.join(
"/media/alpha4TB/ziqi/Parking/Ps_locate_dataset/point", i.strip('.jpg')+'_OA.txt')
# print(fina_path)
ps2, _ = read_pslot(annt_path2)
tmp = output_pic(img_path, output_path,
trans_inv, fina_path, ps2, pix, point_path)
accuracy += tmp
return accuracy
if __name__ == "__main__":
data_dir = '/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/pic'
label_dir = '/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/annotation'
crop_dir = '/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/crop_img'
perspective_dir = '/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/perspective_img'
txt_dir = '/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/point'
cnt = 0
f1 = open(
"/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/train_list.txt", "w")
# f2 = open("./Ps_locate_dataset/val_list.txt", "w")
test_dir = "/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/test_img"
trans_path = "/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/trans_inv"
save_path1 = "/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/src_img"
save_path2 = "/media/home_bak/ziqi/park/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/perspective2_img"
pbar = tqdm(total=len(os.listdir(data_dir)))
for i in os.listdir(data_dir):
# print(i)
annt_file = os.path.join(label_dir, i.strip('.jpg')+'_OA.txt')
img_path = os.path.join(data_dir, i)
ps, _ = read_pslot(annt_file)
Crop_pic(ps, img_path, crop_dir,
perspective_dir, txt_dir, i, trans_path, save_path1, save_path2)
pbar.update(1)
pbar.close()
# acc = []
# for k in range(31):
# print("k", k)
# x1 = output(k)
# x1 = 100 * x1 / 743
# acc.append(x1)
# x1 = round(x1, 3)
# print(acc)
# print(len(acc))
# # 设置画布大小
# plt.figure(figsize=(30, 15))
# # 标题
# plt.title("accruracy distribution")
# # 数据
# plt.bar(range(len(acc)), acc)
# # 横坐标描述
# plt.xlabel('pixel')
# # 纵坐标描述
# plt.ylabel('accuracy')
# # # 设置数字标签
# # for a, b in zip(x, acc):
# # plt.text(a, b, b, ha='center', va='bottom', fontsize=10)
# plt.savefig(
# "/media/alpha4TB/ziqi/Parking/Ps_locate_dataset/PLD_BirdView_TrainingDaraSet_All/accuracy.png")
# 保存训练数据的文件名
filenames = os.listdir(perspective_dir)
filenames.sort()
print(filenames[0])
for i in os.listdir(perspective_dir):
perspective_path = os.path.join(perspective_dir, i)
f1.write(perspective_path)
f1.write('\n')
f1.close()
| 33.156156 | 116 | 0.580926 | [
"Apache-2.0"
] | ziqi123/AutoParking | preprocessing/make_dataset_new.py | 11,369 | Python |
# Generated by the protocol buffer compiler. DO NOT EDIT!
# source: google/cloud/vision_v1p1beta1/proto/image_annotator.proto
import sys
_b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1'))
from google.protobuf.internal import enum_type_wrapper
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
from google.protobuf import descriptor_pb2
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2
from google.cloud.vision_v1p1beta1.proto import geometry_pb2 as google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2
from google.cloud.vision_v1p1beta1.proto import text_annotation_pb2 as google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_text__annotation__pb2
from google.cloud.vision_v1p1beta1.proto import web_detection_pb2 as google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_web__detection__pb2
from google.rpc import status_pb2 as google_dot_rpc_dot_status__pb2
from google.type import color_pb2 as google_dot_type_dot_color__pb2
from google.type import latlng_pb2 as google_dot_type_dot_latlng__pb2
DESCRIPTOR = _descriptor.FileDescriptor(
name='google/cloud/vision_v1p1beta1/proto/image_annotator.proto',
package='google.cloud.vision.v1p1beta1',
syntax='proto3',
serialized_pb=_b('\n9google/cloud/vision_v1p1beta1/proto/image_annotator.proto\x12\x1dgoogle.cloud.vision.v1p1beta1\x1a\x1cgoogle/api/annotations.proto\x1a\x32google/cloud/vision_v1p1beta1/proto/geometry.proto\x1a\x39google/cloud/vision_v1p1beta1/proto/text_annotation.proto\x1a\x37google/cloud/vision_v1p1beta1/proto/web_detection.proto\x1a\x17google/rpc/status.proto\x1a\x17google/type/color.proto\x1a\x18google/type/latlng.proto\"\xe1\x02\n\x07\x46\x65\x61ture\x12\x39\n\x04type\x18\x01 \x01(\x0e\x32+.google.cloud.vision.v1p1beta1.Feature.Type\x12\x13\n\x0bmax_results\x18\x02 \x01(\x05\x12\r\n\x05model\x18\x03 \x01(\t\"\xf6\x01\n\x04Type\x12\x14\n\x10TYPE_UNSPECIFIED\x10\x00\x12\x12\n\x0e\x46\x41\x43\x45_DETECTION\x10\x01\x12\x16\n\x12LANDMARK_DETECTION\x10\x02\x12\x12\n\x0eLOGO_DETECTION\x10\x03\x12\x13\n\x0fLABEL_DETECTION\x10\x04\x12\x12\n\x0eTEXT_DETECTION\x10\x05\x12\x1b\n\x17\x44OCUMENT_TEXT_DETECTION\x10\x0b\x12\x19\n\x15SAFE_SEARCH_DETECTION\x10\x06\x12\x14\n\x10IMAGE_PROPERTIES\x10\x07\x12\x0e\n\nCROP_HINTS\x10\t\x12\x11\n\rWEB_DETECTION\x10\n\"7\n\x0bImageSource\x12\x15\n\rgcs_image_uri\x18\x01 \x01(\t\x12\x11\n\timage_uri\x18\x02 \x01(\t\"T\n\x05Image\x12\x0f\n\x07\x63ontent\x18\x01 \x01(\x0c\x12:\n\x06source\x18\x02 \x01(\x0b\x32*.google.cloud.vision.v1p1beta1.ImageSource\"\x9b\x0e\n\x0e\x46\x61\x63\x65\x41nnotation\x12\x42\n\rbounding_poly\x18\x01 \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.BoundingPoly\x12\x45\n\x10\x66\x64_bounding_poly\x18\x02 \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.BoundingPoly\x12I\n\tlandmarks\x18\x03 \x03(\x0b\x32\x36.google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark\x12\x12\n\nroll_angle\x18\x04 \x01(\x02\x12\x11\n\tpan_angle\x18\x05 \x01(\x02\x12\x12\n\ntilt_angle\x18\x06 \x01(\x02\x12\x1c\n\x14\x64\x65tection_confidence\x18\x07 \x01(\x02\x12\x1e\n\x16landmarking_confidence\x18\x08 \x01(\x02\x12\x41\n\x0ejoy_likelihood\x18\t \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x44\n\x11sorrow_likelihood\x18\n \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x43\n\x10\x61nger_likelihood\x18\x0b \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x46\n\x13surprise_likelihood\x18\x0c \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12K\n\x18under_exposed_likelihood\x18\r \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x45\n\x12\x62lurred_likelihood\x18\x0e \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x46\n\x13headwear_likelihood\x18\x0f \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x1a\xc7\x07\n\x08Landmark\x12I\n\x04type\x18\x03 \x01(\x0e\x32;.google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark.Type\x12\x39\n\x08position\x18\x04 \x01(\x0b\x32\'.google.cloud.vision.v1p1beta1.Position\"\xb4\x06\n\x04Type\x12\x14\n\x10UNKNOWN_LANDMARK\x10\x00\x12\x0c\n\x08LEFT_EYE\x10\x01\x12\r\n\tRIGHT_EYE\x10\x02\x12\x18\n\x14LEFT_OF_LEFT_EYEBROW\x10\x03\x12\x19\n\x15RIGHT_OF_LEFT_EYEBROW\x10\x04\x12\x19\n\x15LEFT_OF_RIGHT_EYEBROW\x10\x05\x12\x1a\n\x16RIGHT_OF_RIGHT_EYEBROW\x10\x06\x12\x19\n\x15MIDPOINT_BETWEEN_EYES\x10\x07\x12\x0c\n\x08NOSE_TIP\x10\x08\x12\r\n\tUPPER_LIP\x10\t\x12\r\n\tLOWER_LIP\x10\n\x12\x0e\n\nMOUTH_LEFT\x10\x0b\x12\x0f\n\x0bMOUTH_RIGHT\x10\x0c\x12\x10\n\x0cMOUTH_CENTER\x10\r\x12\x15\n\x11NOSE_BOTTOM_RIGHT\x10\x0e\x12\x14\n\x10NOSE_BOTTOM_LEFT\x10\x0f\x12\x16\n\x12NOSE_BOTTOM_CENTER\x10\x10\x12\x19\n\x15LEFT_EYE_TOP_BOUNDARY\x10\x11\x12\x19\n\x15LEFT_EYE_RIGHT_CORNER\x10\x12\x12\x1c\n\x18LEFT_EYE_BOTTOM_BOUNDARY\x10\x13\x12\x18\n\x14LEFT_EYE_LEFT_CORNER\x10\x14\x12\x1a\n\x16RIGHT_EYE_TOP_BOUNDARY\x10\x15\x12\x1a\n\x16RIGHT_EYE_RIGHT_CORNER\x10\x16\x12\x1d\n\x19RIGHT_EYE_BOTTOM_BOUNDARY\x10\x17\x12\x19\n\x15RIGHT_EYE_LEFT_CORNER\x10\x18\x12\x1f\n\x1bLEFT_EYEBROW_UPPER_MIDPOINT\x10\x19\x12 \n\x1cRIGHT_EYEBROW_UPPER_MIDPOINT\x10\x1a\x12\x14\n\x10LEFT_EAR_TRAGION\x10\x1b\x12\x15\n\x11RIGHT_EAR_TRAGION\x10\x1c\x12\x12\n\x0eLEFT_EYE_PUPIL\x10\x1d\x12\x13\n\x0fRIGHT_EYE_PUPIL\x10\x1e\x12\x15\n\x11\x46OREHEAD_GLABELLA\x10\x1f\x12\x11\n\rCHIN_GNATHION\x10 \x12\x14\n\x10\x43HIN_LEFT_GONION\x10!\x12\x15\n\x11\x43HIN_RIGHT_GONION\x10\"\"4\n\x0cLocationInfo\x12$\n\x07lat_lng\x18\x01 \x01(\x0b\x32\x13.google.type.LatLng\"=\n\x08Property\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\r\n\x05value\x18\x02 \x01(\t\x12\x14\n\x0cuint64_value\x18\x03 \x01(\x04\"\xbc\x02\n\x10\x45ntityAnnotation\x12\x0b\n\x03mid\x18\x01 \x01(\t\x12\x0e\n\x06locale\x18\x02 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x03 \x01(\t\x12\r\n\x05score\x18\x04 \x01(\x02\x12\x12\n\nconfidence\x18\x05 \x01(\x02\x12\x12\n\ntopicality\x18\x06 \x01(\x02\x12\x42\n\rbounding_poly\x18\x07 \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.BoundingPoly\x12>\n\tlocations\x18\x08 \x03(\x0b\x32+.google.cloud.vision.v1p1beta1.LocationInfo\x12;\n\nproperties\x18\t \x03(\x0b\x32\'.google.cloud.vision.v1p1beta1.Property\"\xbc\x02\n\x14SafeSearchAnnotation\x12\x38\n\x05\x61\x64ult\x18\x01 \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x38\n\x05spoof\x18\x02 \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12:\n\x07medical\x18\x03 \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12;\n\x08violence\x18\x04 \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\x12\x37\n\x04racy\x18\t \x01(\x0e\x32).google.cloud.vision.v1p1beta1.Likelihood\"a\n\x0bLatLongRect\x12(\n\x0bmin_lat_lng\x18\x01 \x01(\x0b\x32\x13.google.type.LatLng\x12(\n\x0bmax_lat_lng\x18\x02 \x01(\x0b\x32\x13.google.type.LatLng\"U\n\tColorInfo\x12!\n\x05\x63olor\x18\x01 \x01(\x0b\x32\x12.google.type.Color\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x16\n\x0epixel_fraction\x18\x03 \x01(\x02\"T\n\x18\x44ominantColorsAnnotation\x12\x38\n\x06\x63olors\x18\x01 \x03(\x0b\x32(.google.cloud.vision.v1p1beta1.ColorInfo\"c\n\x0fImageProperties\x12P\n\x0f\x64ominant_colors\x18\x01 \x01(\x0b\x32\x37.google.cloud.vision.v1p1beta1.DominantColorsAnnotation\"\x7f\n\x08\x43ropHint\x12\x42\n\rbounding_poly\x18\x01 \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.BoundingPoly\x12\x12\n\nconfidence\x18\x02 \x01(\x02\x12\x1b\n\x13importance_fraction\x18\x03 \x01(\x02\"R\n\x13\x43ropHintsAnnotation\x12;\n\ncrop_hints\x18\x01 \x03(\x0b\x32\'.google.cloud.vision.v1p1beta1.CropHint\"(\n\x0f\x43ropHintsParams\x12\x15\n\raspect_ratios\x18\x01 \x03(\x02\"1\n\x12WebDetectionParams\x12\x1b\n\x13include_geo_results\x18\x02 \x01(\x08\"\x85\x02\n\x0cImageContext\x12\x41\n\rlat_long_rect\x18\x01 \x01(\x0b\x32*.google.cloud.vision.v1p1beta1.LatLongRect\x12\x16\n\x0elanguage_hints\x18\x02 \x03(\t\x12I\n\x11\x63rop_hints_params\x18\x04 \x01(\x0b\x32..google.cloud.vision.v1p1beta1.CropHintsParams\x12O\n\x14web_detection_params\x18\x06 \x01(\x0b\x32\x31.google.cloud.vision.v1p1beta1.WebDetectionParams\"\xc9\x01\n\x14\x41nnotateImageRequest\x12\x33\n\x05image\x18\x01 \x01(\x0b\x32$.google.cloud.vision.v1p1beta1.Image\x12\x38\n\x08\x66\x65\x61tures\x18\x02 \x03(\x0b\x32&.google.cloud.vision.v1p1beta1.Feature\x12\x42\n\rimage_context\x18\x03 \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.ImageContext\"\xc2\x06\n\x15\x41nnotateImageResponse\x12G\n\x10\x66\x61\x63\x65_annotations\x18\x01 \x03(\x0b\x32-.google.cloud.vision.v1p1beta1.FaceAnnotation\x12M\n\x14landmark_annotations\x18\x02 \x03(\x0b\x32/.google.cloud.vision.v1p1beta1.EntityAnnotation\x12I\n\x10logo_annotations\x18\x03 \x03(\x0b\x32/.google.cloud.vision.v1p1beta1.EntityAnnotation\x12J\n\x11label_annotations\x18\x04 \x03(\x0b\x32/.google.cloud.vision.v1p1beta1.EntityAnnotation\x12I\n\x10text_annotations\x18\x05 \x03(\x0b\x32/.google.cloud.vision.v1p1beta1.EntityAnnotation\x12K\n\x14\x66ull_text_annotation\x18\x0c \x01(\x0b\x32-.google.cloud.vision.v1p1beta1.TextAnnotation\x12S\n\x16safe_search_annotation\x18\x06 \x01(\x0b\x32\x33.google.cloud.vision.v1p1beta1.SafeSearchAnnotation\x12S\n\x1bimage_properties_annotation\x18\x08 \x01(\x0b\x32..google.cloud.vision.v1p1beta1.ImageProperties\x12Q\n\x15\x63rop_hints_annotation\x18\x0b \x01(\x0b\x32\x32.google.cloud.vision.v1p1beta1.CropHintsAnnotation\x12\x42\n\rweb_detection\x18\r \x01(\x0b\x32+.google.cloud.vision.v1p1beta1.WebDetection\x12!\n\x05\x65rror\x18\t \x01(\x0b\x32\x12.google.rpc.Status\"c\n\x1a\x42\x61tchAnnotateImagesRequest\x12\x45\n\x08requests\x18\x01 \x03(\x0b\x32\x33.google.cloud.vision.v1p1beta1.AnnotateImageRequest\"f\n\x1b\x42\x61tchAnnotateImagesResponse\x12G\n\tresponses\x18\x01 \x03(\x0b\x32\x34.google.cloud.vision.v1p1beta1.AnnotateImageResponse*e\n\nLikelihood\x12\x0b\n\x07UNKNOWN\x10\x00\x12\x11\n\rVERY_UNLIKELY\x10\x01\x12\x0c\n\x08UNLIKELY\x10\x02\x12\x0c\n\x08POSSIBLE\x10\x03\x12\n\n\x06LIKELY\x10\x04\x12\x0f\n\x0bVERY_LIKELY\x10\x05\x32\xc6\x01\n\x0eImageAnnotator\x12\xb3\x01\n\x13\x42\x61tchAnnotateImages\x12\x39.google.cloud.vision.v1p1beta1.BatchAnnotateImagesRequest\x1a:.google.cloud.vision.v1p1beta1.BatchAnnotateImagesResponse\"%\x82\xd3\xe4\x93\x02\x1f\"\x1a/v1p1beta1/images:annotate:\x01*B\x82\x01\n!com.google.cloud.vision.v1p1beta1B\x13ImageAnnotatorProtoP\x01ZCgoogle.golang.org/genproto/googleapis/cloud/vision/v1p1beta1;vision\xf8\x01\x01\x62\x06proto3')
,
dependencies=[google_dot_api_dot_annotations__pb2.DESCRIPTOR,google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2.DESCRIPTOR,google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_text__annotation__pb2.DESCRIPTOR,google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_web__detection__pb2.DESCRIPTOR,google_dot_rpc_dot_status__pb2.DESCRIPTOR,google_dot_type_dot_color__pb2.DESCRIPTOR,google_dot_type_dot_latlng__pb2.DESCRIPTOR,])
_sym_db.RegisterFileDescriptor(DESCRIPTOR)
_LIKELIHOOD = _descriptor.EnumDescriptor(
name='Likelihood',
full_name='google.cloud.vision.v1p1beta1.Likelihood',
filename=None,
file=DESCRIPTOR,
values=[
_descriptor.EnumValueDescriptor(
name='UNKNOWN', index=0, number=0,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='VERY_UNLIKELY', index=1, number=1,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='UNLIKELY', index=2, number=2,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='POSSIBLE', index=3, number=3,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LIKELY', index=4, number=4,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='VERY_LIKELY', index=5, number=5,
options=None,
type=None),
],
containing_type=None,
options=None,
serialized_start=5631,
serialized_end=5732,
)
_sym_db.RegisterEnumDescriptor(_LIKELIHOOD)
Likelihood = enum_type_wrapper.EnumTypeWrapper(_LIKELIHOOD)
UNKNOWN = 0
VERY_UNLIKELY = 1
UNLIKELY = 2
POSSIBLE = 3
LIKELY = 4
VERY_LIKELY = 5
_FEATURE_TYPE = _descriptor.EnumDescriptor(
name='Type',
full_name='google.cloud.vision.v1p1beta1.Feature.Type',
filename=None,
file=DESCRIPTOR,
values=[
_descriptor.EnumValueDescriptor(
name='TYPE_UNSPECIFIED', index=0, number=0,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='FACE_DETECTION', index=1, number=1,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LANDMARK_DETECTION', index=2, number=2,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LOGO_DETECTION', index=3, number=3,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LABEL_DETECTION', index=4, number=4,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='TEXT_DETECTION', index=5, number=5,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='DOCUMENT_TEXT_DETECTION', index=6, number=11,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='SAFE_SEARCH_DETECTION', index=7, number=6,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='IMAGE_PROPERTIES', index=8, number=7,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='CROP_HINTS', index=9, number=9,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='WEB_DETECTION', index=10, number=10,
options=None,
type=None),
],
containing_type=None,
options=None,
serialized_start=474,
serialized_end=720,
)
_sym_db.RegisterEnumDescriptor(_FEATURE_TYPE)
_FACEANNOTATION_LANDMARK_TYPE = _descriptor.EnumDescriptor(
name='Type',
full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark.Type',
filename=None,
file=DESCRIPTOR,
values=[
_descriptor.EnumValueDescriptor(
name='UNKNOWN_LANDMARK', index=0, number=0,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE', index=1, number=1,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE', index=2, number=2,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_OF_LEFT_EYEBROW', index=3, number=3,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_OF_LEFT_EYEBROW', index=4, number=4,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_OF_RIGHT_EYEBROW', index=5, number=5,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_OF_RIGHT_EYEBROW', index=6, number=6,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='MIDPOINT_BETWEEN_EYES', index=7, number=7,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='NOSE_TIP', index=8, number=8,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='UPPER_LIP', index=9, number=9,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LOWER_LIP', index=10, number=10,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='MOUTH_LEFT', index=11, number=11,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='MOUTH_RIGHT', index=12, number=12,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='MOUTH_CENTER', index=13, number=13,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='NOSE_BOTTOM_RIGHT', index=14, number=14,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='NOSE_BOTTOM_LEFT', index=15, number=15,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='NOSE_BOTTOM_CENTER', index=16, number=16,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE_TOP_BOUNDARY', index=17, number=17,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE_RIGHT_CORNER', index=18, number=18,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE_BOTTOM_BOUNDARY', index=19, number=19,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE_LEFT_CORNER', index=20, number=20,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE_TOP_BOUNDARY', index=21, number=21,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE_RIGHT_CORNER', index=22, number=22,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE_BOTTOM_BOUNDARY', index=23, number=23,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE_LEFT_CORNER', index=24, number=24,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYEBROW_UPPER_MIDPOINT', index=25, number=25,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYEBROW_UPPER_MIDPOINT', index=26, number=26,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EAR_TRAGION', index=27, number=27,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EAR_TRAGION', index=28, number=28,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='LEFT_EYE_PUPIL', index=29, number=29,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='RIGHT_EYE_PUPIL', index=30, number=30,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='FOREHEAD_GLABELLA', index=31, number=31,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='CHIN_GNATHION', index=32, number=32,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='CHIN_LEFT_GONION', index=33, number=33,
options=None,
type=None),
_descriptor.EnumValueDescriptor(
name='CHIN_RIGHT_GONION', index=34, number=34,
options=None,
type=None),
],
containing_type=None,
options=None,
serialized_start=1865,
serialized_end=2685,
)
_sym_db.RegisterEnumDescriptor(_FACEANNOTATION_LANDMARK_TYPE)
_FEATURE = _descriptor.Descriptor(
name='Feature',
full_name='google.cloud.vision.v1p1beta1.Feature',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='type', full_name='google.cloud.vision.v1p1beta1.Feature.type', index=0,
number=1, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='max_results', full_name='google.cloud.vision.v1p1beta1.Feature.max_results', index=1,
number=2, type=5, cpp_type=1, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='model', full_name='google.cloud.vision.v1p1beta1.Feature.model', index=2,
number=3, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
_FEATURE_TYPE,
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=367,
serialized_end=720,
)
_IMAGESOURCE = _descriptor.Descriptor(
name='ImageSource',
full_name='google.cloud.vision.v1p1beta1.ImageSource',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='gcs_image_uri', full_name='google.cloud.vision.v1p1beta1.ImageSource.gcs_image_uri', index=0,
number=1, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='image_uri', full_name='google.cloud.vision.v1p1beta1.ImageSource.image_uri', index=1,
number=2, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=722,
serialized_end=777,
)
_IMAGE = _descriptor.Descriptor(
name='Image',
full_name='google.cloud.vision.v1p1beta1.Image',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='content', full_name='google.cloud.vision.v1p1beta1.Image.content', index=0,
number=1, type=12, cpp_type=9, label=1,
has_default_value=False, default_value=_b(""),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='source', full_name='google.cloud.vision.v1p1beta1.Image.source', index=1,
number=2, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=779,
serialized_end=863,
)
_FACEANNOTATION_LANDMARK = _descriptor.Descriptor(
name='Landmark',
full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='type', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark.type', index=0,
number=3, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='position', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark.position', index=1,
number=4, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
_FACEANNOTATION_LANDMARK_TYPE,
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=1718,
serialized_end=2685,
)
_FACEANNOTATION = _descriptor.Descriptor(
name='FaceAnnotation',
full_name='google.cloud.vision.v1p1beta1.FaceAnnotation',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='bounding_poly', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.bounding_poly', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='fd_bounding_poly', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.fd_bounding_poly', index=1,
number=2, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='landmarks', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.landmarks', index=2,
number=3, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='roll_angle', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.roll_angle', index=3,
number=4, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='pan_angle', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.pan_angle', index=4,
number=5, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='tilt_angle', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.tilt_angle', index=5,
number=6, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='detection_confidence', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.detection_confidence', index=6,
number=7, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='landmarking_confidence', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.landmarking_confidence', index=7,
number=8, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='joy_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.joy_likelihood', index=8,
number=9, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='sorrow_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.sorrow_likelihood', index=9,
number=10, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='anger_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.anger_likelihood', index=10,
number=11, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='surprise_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.surprise_likelihood', index=11,
number=12, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='under_exposed_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.under_exposed_likelihood', index=12,
number=13, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='blurred_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.blurred_likelihood', index=13,
number=14, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='headwear_likelihood', full_name='google.cloud.vision.v1p1beta1.FaceAnnotation.headwear_likelihood', index=14,
number=15, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[_FACEANNOTATION_LANDMARK, ],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=866,
serialized_end=2685,
)
_LOCATIONINFO = _descriptor.Descriptor(
name='LocationInfo',
full_name='google.cloud.vision.v1p1beta1.LocationInfo',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='lat_lng', full_name='google.cloud.vision.v1p1beta1.LocationInfo.lat_lng', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=2687,
serialized_end=2739,
)
_PROPERTY = _descriptor.Descriptor(
name='Property',
full_name='google.cloud.vision.v1p1beta1.Property',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='name', full_name='google.cloud.vision.v1p1beta1.Property.name', index=0,
number=1, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='value', full_name='google.cloud.vision.v1p1beta1.Property.value', index=1,
number=2, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='uint64_value', full_name='google.cloud.vision.v1p1beta1.Property.uint64_value', index=2,
number=3, type=4, cpp_type=4, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=2741,
serialized_end=2802,
)
_ENTITYANNOTATION = _descriptor.Descriptor(
name='EntityAnnotation',
full_name='google.cloud.vision.v1p1beta1.EntityAnnotation',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='mid', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.mid', index=0,
number=1, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='locale', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.locale', index=1,
number=2, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='description', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.description', index=2,
number=3, type=9, cpp_type=9, label=1,
has_default_value=False, default_value=_b("").decode('utf-8'),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='score', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.score', index=3,
number=4, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='confidence', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.confidence', index=4,
number=5, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='topicality', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.topicality', index=5,
number=6, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='bounding_poly', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.bounding_poly', index=6,
number=7, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='locations', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.locations', index=7,
number=8, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='properties', full_name='google.cloud.vision.v1p1beta1.EntityAnnotation.properties', index=8,
number=9, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=2805,
serialized_end=3121,
)
_SAFESEARCHANNOTATION = _descriptor.Descriptor(
name='SafeSearchAnnotation',
full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='adult', full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation.adult', index=0,
number=1, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='spoof', full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation.spoof', index=1,
number=2, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='medical', full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation.medical', index=2,
number=3, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='violence', full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation.violence', index=3,
number=4, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='racy', full_name='google.cloud.vision.v1p1beta1.SafeSearchAnnotation.racy', index=4,
number=9, type=14, cpp_type=8, label=1,
has_default_value=False, default_value=0,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3124,
serialized_end=3440,
)
_LATLONGRECT = _descriptor.Descriptor(
name='LatLongRect',
full_name='google.cloud.vision.v1p1beta1.LatLongRect',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='min_lat_lng', full_name='google.cloud.vision.v1p1beta1.LatLongRect.min_lat_lng', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='max_lat_lng', full_name='google.cloud.vision.v1p1beta1.LatLongRect.max_lat_lng', index=1,
number=2, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3442,
serialized_end=3539,
)
_COLORINFO = _descriptor.Descriptor(
name='ColorInfo',
full_name='google.cloud.vision.v1p1beta1.ColorInfo',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='color', full_name='google.cloud.vision.v1p1beta1.ColorInfo.color', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='score', full_name='google.cloud.vision.v1p1beta1.ColorInfo.score', index=1,
number=2, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='pixel_fraction', full_name='google.cloud.vision.v1p1beta1.ColorInfo.pixel_fraction', index=2,
number=3, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3541,
serialized_end=3626,
)
_DOMINANTCOLORSANNOTATION = _descriptor.Descriptor(
name='DominantColorsAnnotation',
full_name='google.cloud.vision.v1p1beta1.DominantColorsAnnotation',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='colors', full_name='google.cloud.vision.v1p1beta1.DominantColorsAnnotation.colors', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3628,
serialized_end=3712,
)
_IMAGEPROPERTIES = _descriptor.Descriptor(
name='ImageProperties',
full_name='google.cloud.vision.v1p1beta1.ImageProperties',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='dominant_colors', full_name='google.cloud.vision.v1p1beta1.ImageProperties.dominant_colors', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3714,
serialized_end=3813,
)
_CROPHINT = _descriptor.Descriptor(
name='CropHint',
full_name='google.cloud.vision.v1p1beta1.CropHint',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='bounding_poly', full_name='google.cloud.vision.v1p1beta1.CropHint.bounding_poly', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='confidence', full_name='google.cloud.vision.v1p1beta1.CropHint.confidence', index=1,
number=2, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='importance_fraction', full_name='google.cloud.vision.v1p1beta1.CropHint.importance_fraction', index=2,
number=3, type=2, cpp_type=6, label=1,
has_default_value=False, default_value=float(0),
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3815,
serialized_end=3942,
)
_CROPHINTSANNOTATION = _descriptor.Descriptor(
name='CropHintsAnnotation',
full_name='google.cloud.vision.v1p1beta1.CropHintsAnnotation',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='crop_hints', full_name='google.cloud.vision.v1p1beta1.CropHintsAnnotation.crop_hints', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=3944,
serialized_end=4026,
)
_CROPHINTSPARAMS = _descriptor.Descriptor(
name='CropHintsParams',
full_name='google.cloud.vision.v1p1beta1.CropHintsParams',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='aspect_ratios', full_name='google.cloud.vision.v1p1beta1.CropHintsParams.aspect_ratios', index=0,
number=1, type=2, cpp_type=6, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=4028,
serialized_end=4068,
)
_WEBDETECTIONPARAMS = _descriptor.Descriptor(
name='WebDetectionParams',
full_name='google.cloud.vision.v1p1beta1.WebDetectionParams',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='include_geo_results', full_name='google.cloud.vision.v1p1beta1.WebDetectionParams.include_geo_results', index=0,
number=2, type=8, cpp_type=7, label=1,
has_default_value=False, default_value=False,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=4070,
serialized_end=4119,
)
_IMAGECONTEXT = _descriptor.Descriptor(
name='ImageContext',
full_name='google.cloud.vision.v1p1beta1.ImageContext',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='lat_long_rect', full_name='google.cloud.vision.v1p1beta1.ImageContext.lat_long_rect', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='language_hints', full_name='google.cloud.vision.v1p1beta1.ImageContext.language_hints', index=1,
number=2, type=9, cpp_type=9, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='crop_hints_params', full_name='google.cloud.vision.v1p1beta1.ImageContext.crop_hints_params', index=2,
number=4, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='web_detection_params', full_name='google.cloud.vision.v1p1beta1.ImageContext.web_detection_params', index=3,
number=6, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=4122,
serialized_end=4383,
)
_ANNOTATEIMAGEREQUEST = _descriptor.Descriptor(
name='AnnotateImageRequest',
full_name='google.cloud.vision.v1p1beta1.AnnotateImageRequest',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='image', full_name='google.cloud.vision.v1p1beta1.AnnotateImageRequest.image', index=0,
number=1, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='features', full_name='google.cloud.vision.v1p1beta1.AnnotateImageRequest.features', index=1,
number=2, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='image_context', full_name='google.cloud.vision.v1p1beta1.AnnotateImageRequest.image_context', index=2,
number=3, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=4386,
serialized_end=4587,
)
_ANNOTATEIMAGERESPONSE = _descriptor.Descriptor(
name='AnnotateImageResponse',
full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='face_annotations', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.face_annotations', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='landmark_annotations', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.landmark_annotations', index=1,
number=2, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='logo_annotations', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.logo_annotations', index=2,
number=3, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='label_annotations', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.label_annotations', index=3,
number=4, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='text_annotations', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.text_annotations', index=4,
number=5, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='full_text_annotation', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.full_text_annotation', index=5,
number=12, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='safe_search_annotation', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.safe_search_annotation', index=6,
number=6, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='image_properties_annotation', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.image_properties_annotation', index=7,
number=8, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='crop_hints_annotation', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.crop_hints_annotation', index=8,
number=11, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='web_detection', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.web_detection', index=9,
number=13, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
_descriptor.FieldDescriptor(
name='error', full_name='google.cloud.vision.v1p1beta1.AnnotateImageResponse.error', index=10,
number=9, type=11, cpp_type=10, label=1,
has_default_value=False, default_value=None,
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=4590,
serialized_end=5424,
)
_BATCHANNOTATEIMAGESREQUEST = _descriptor.Descriptor(
name='BatchAnnotateImagesRequest',
full_name='google.cloud.vision.v1p1beta1.BatchAnnotateImagesRequest',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='requests', full_name='google.cloud.vision.v1p1beta1.BatchAnnotateImagesRequest.requests', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=5426,
serialized_end=5525,
)
_BATCHANNOTATEIMAGESRESPONSE = _descriptor.Descriptor(
name='BatchAnnotateImagesResponse',
full_name='google.cloud.vision.v1p1beta1.BatchAnnotateImagesResponse',
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name='responses', full_name='google.cloud.vision.v1p1beta1.BatchAnnotateImagesResponse.responses', index=0,
number=1, type=11, cpp_type=10, label=3,
has_default_value=False, default_value=[],
message_type=None, enum_type=None, containing_type=None,
is_extension=False, extension_scope=None,
options=None),
],
extensions=[
],
nested_types=[],
enum_types=[
],
options=None,
is_extendable=False,
syntax='proto3',
extension_ranges=[],
oneofs=[
],
serialized_start=5527,
serialized_end=5629,
)
_FEATURE.fields_by_name['type'].enum_type = _FEATURE_TYPE
_FEATURE_TYPE.containing_type = _FEATURE
_IMAGE.fields_by_name['source'].message_type = _IMAGESOURCE
_FACEANNOTATION_LANDMARK.fields_by_name['type'].enum_type = _FACEANNOTATION_LANDMARK_TYPE
_FACEANNOTATION_LANDMARK.fields_by_name['position'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2._POSITION
_FACEANNOTATION_LANDMARK.containing_type = _FACEANNOTATION
_FACEANNOTATION_LANDMARK_TYPE.containing_type = _FACEANNOTATION_LANDMARK
_FACEANNOTATION.fields_by_name['bounding_poly'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2._BOUNDINGPOLY
_FACEANNOTATION.fields_by_name['fd_bounding_poly'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2._BOUNDINGPOLY
_FACEANNOTATION.fields_by_name['landmarks'].message_type = _FACEANNOTATION_LANDMARK
_FACEANNOTATION.fields_by_name['joy_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['sorrow_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['anger_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['surprise_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['under_exposed_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['blurred_likelihood'].enum_type = _LIKELIHOOD
_FACEANNOTATION.fields_by_name['headwear_likelihood'].enum_type = _LIKELIHOOD
_LOCATIONINFO.fields_by_name['lat_lng'].message_type = google_dot_type_dot_latlng__pb2._LATLNG
_ENTITYANNOTATION.fields_by_name['bounding_poly'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2._BOUNDINGPOLY
_ENTITYANNOTATION.fields_by_name['locations'].message_type = _LOCATIONINFO
_ENTITYANNOTATION.fields_by_name['properties'].message_type = _PROPERTY
_SAFESEARCHANNOTATION.fields_by_name['adult'].enum_type = _LIKELIHOOD
_SAFESEARCHANNOTATION.fields_by_name['spoof'].enum_type = _LIKELIHOOD
_SAFESEARCHANNOTATION.fields_by_name['medical'].enum_type = _LIKELIHOOD
_SAFESEARCHANNOTATION.fields_by_name['violence'].enum_type = _LIKELIHOOD
_SAFESEARCHANNOTATION.fields_by_name['racy'].enum_type = _LIKELIHOOD
_LATLONGRECT.fields_by_name['min_lat_lng'].message_type = google_dot_type_dot_latlng__pb2._LATLNG
_LATLONGRECT.fields_by_name['max_lat_lng'].message_type = google_dot_type_dot_latlng__pb2._LATLNG
_COLORINFO.fields_by_name['color'].message_type = google_dot_type_dot_color__pb2._COLOR
_DOMINANTCOLORSANNOTATION.fields_by_name['colors'].message_type = _COLORINFO
_IMAGEPROPERTIES.fields_by_name['dominant_colors'].message_type = _DOMINANTCOLORSANNOTATION
_CROPHINT.fields_by_name['bounding_poly'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_geometry__pb2._BOUNDINGPOLY
_CROPHINTSANNOTATION.fields_by_name['crop_hints'].message_type = _CROPHINT
_IMAGECONTEXT.fields_by_name['lat_long_rect'].message_type = _LATLONGRECT
_IMAGECONTEXT.fields_by_name['crop_hints_params'].message_type = _CROPHINTSPARAMS
_IMAGECONTEXT.fields_by_name['web_detection_params'].message_type = _WEBDETECTIONPARAMS
_ANNOTATEIMAGEREQUEST.fields_by_name['image'].message_type = _IMAGE
_ANNOTATEIMAGEREQUEST.fields_by_name['features'].message_type = _FEATURE
_ANNOTATEIMAGEREQUEST.fields_by_name['image_context'].message_type = _IMAGECONTEXT
_ANNOTATEIMAGERESPONSE.fields_by_name['face_annotations'].message_type = _FACEANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['landmark_annotations'].message_type = _ENTITYANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['logo_annotations'].message_type = _ENTITYANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['label_annotations'].message_type = _ENTITYANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['text_annotations'].message_type = _ENTITYANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['full_text_annotation'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_text__annotation__pb2._TEXTANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['safe_search_annotation'].message_type = _SAFESEARCHANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['image_properties_annotation'].message_type = _IMAGEPROPERTIES
_ANNOTATEIMAGERESPONSE.fields_by_name['crop_hints_annotation'].message_type = _CROPHINTSANNOTATION
_ANNOTATEIMAGERESPONSE.fields_by_name['web_detection'].message_type = google_dot_cloud_dot_vision__v1p1beta1_dot_proto_dot_web__detection__pb2._WEBDETECTION
_ANNOTATEIMAGERESPONSE.fields_by_name['error'].message_type = google_dot_rpc_dot_status__pb2._STATUS
_BATCHANNOTATEIMAGESREQUEST.fields_by_name['requests'].message_type = _ANNOTATEIMAGEREQUEST
_BATCHANNOTATEIMAGESRESPONSE.fields_by_name['responses'].message_type = _ANNOTATEIMAGERESPONSE
DESCRIPTOR.message_types_by_name['Feature'] = _FEATURE
DESCRIPTOR.message_types_by_name['ImageSource'] = _IMAGESOURCE
DESCRIPTOR.message_types_by_name['Image'] = _IMAGE
DESCRIPTOR.message_types_by_name['FaceAnnotation'] = _FACEANNOTATION
DESCRIPTOR.message_types_by_name['LocationInfo'] = _LOCATIONINFO
DESCRIPTOR.message_types_by_name['Property'] = _PROPERTY
DESCRIPTOR.message_types_by_name['EntityAnnotation'] = _ENTITYANNOTATION
DESCRIPTOR.message_types_by_name['SafeSearchAnnotation'] = _SAFESEARCHANNOTATION
DESCRIPTOR.message_types_by_name['LatLongRect'] = _LATLONGRECT
DESCRIPTOR.message_types_by_name['ColorInfo'] = _COLORINFO
DESCRIPTOR.message_types_by_name['DominantColorsAnnotation'] = _DOMINANTCOLORSANNOTATION
DESCRIPTOR.message_types_by_name['ImageProperties'] = _IMAGEPROPERTIES
DESCRIPTOR.message_types_by_name['CropHint'] = _CROPHINT
DESCRIPTOR.message_types_by_name['CropHintsAnnotation'] = _CROPHINTSANNOTATION
DESCRIPTOR.message_types_by_name['CropHintsParams'] = _CROPHINTSPARAMS
DESCRIPTOR.message_types_by_name['WebDetectionParams'] = _WEBDETECTIONPARAMS
DESCRIPTOR.message_types_by_name['ImageContext'] = _IMAGECONTEXT
DESCRIPTOR.message_types_by_name['AnnotateImageRequest'] = _ANNOTATEIMAGEREQUEST
DESCRIPTOR.message_types_by_name['AnnotateImageResponse'] = _ANNOTATEIMAGERESPONSE
DESCRIPTOR.message_types_by_name['BatchAnnotateImagesRequest'] = _BATCHANNOTATEIMAGESREQUEST
DESCRIPTOR.message_types_by_name['BatchAnnotateImagesResponse'] = _BATCHANNOTATEIMAGESRESPONSE
DESCRIPTOR.enum_types_by_name['Likelihood'] = _LIKELIHOOD
Feature = _reflection.GeneratedProtocolMessageType('Feature', (_message.Message,), dict(
DESCRIPTOR = _FEATURE,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Users describe the type of Google Cloud Vision API tasks to perform over
images by using *Feature*\ s. Each Feature indicates a type of image
detection task to perform. Features encode the Cloud Vision API vertical
to operate on and the number of top-scoring results to return.
Attributes:
type:
The feature type.
max_results:
Maximum number of results of this type.
model:
Model to use for the feature. Supported values:
"builtin/stable" (the default if unset) and "builtin/latest".
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.Feature)
))
_sym_db.RegisterMessage(Feature)
ImageSource = _reflection.GeneratedProtocolMessageType('ImageSource', (_message.Message,), dict(
DESCRIPTOR = _IMAGESOURCE,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """External image source (Google Cloud Storage image location).
Attributes:
gcs_image_uri:
NOTE: For new code ``image_uri`` below is preferred. Google
Cloud Storage image URI, which must be in the following form:
``gs://bucket_name/object_name`` (for details, see `Google
Cloud Storage Request URIs
<https://cloud.google.com/storage/docs/reference-uris>`__).
NOTE: Cloud Storage object versioning is not supported.
image_uri:
Image URI which supports: 1) Google Cloud Storage image URI,
which must be in the following form:
``gs://bucket_name/object_name`` (for details, see `Google
Cloud Storage Request URIs
<https://cloud.google.com/storage/docs/reference-uris>`__).
NOTE: Cloud Storage object versioning is not supported. 2)
Publicly accessible image HTTP/HTTPS URL. This is preferred
over the legacy ``gcs_image_uri`` above. When both
``gcs_image_uri`` and ``image_uri`` are specified,
``image_uri`` takes precedence.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.ImageSource)
))
_sym_db.RegisterMessage(ImageSource)
Image = _reflection.GeneratedProtocolMessageType('Image', (_message.Message,), dict(
DESCRIPTOR = _IMAGE,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Client image to perform Google Cloud Vision API tasks over.
Attributes:
content:
Image content, represented as a stream of bytes. Note: as with
all ``bytes`` fields, protobuffers use a pure binary
representation, whereas JSON representations use base64.
source:
Google Cloud Storage image location. If both ``content`` and
``source`` are provided for an image, ``content`` takes
precedence and is used to perform the image annotation
request.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.Image)
))
_sym_db.RegisterMessage(Image)
FaceAnnotation = _reflection.GeneratedProtocolMessageType('FaceAnnotation', (_message.Message,), dict(
Landmark = _reflection.GeneratedProtocolMessageType('Landmark', (_message.Message,), dict(
DESCRIPTOR = _FACEANNOTATION_LANDMARK,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """A face-specific landmark (for example, a face feature).
Attributes:
type:
Face landmark type.
position:
Face landmark position.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.FaceAnnotation.Landmark)
))
,
DESCRIPTOR = _FACEANNOTATION,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """A face annotation object contains the results of face detection.
Attributes:
bounding_poly:
The bounding polygon around the face. The coordinates of the
bounding box are in the original image's scale, as returned in
``ImageParams``. The bounding box is computed to "frame" the
face in accordance with human expectations. It is based on the
landmarker results. Note that one or more x and/or y
coordinates may not be generated in the ``BoundingPoly`` (the
polygon will be unbounded) if only a partial face appears in
the image to be annotated.
fd_bounding_poly:
The ``fd_bounding_poly`` bounding polygon is tighter than the
``boundingPoly``, and encloses only the skin part of the face.
Typically, it is used to eliminate the face from any image
analysis that detects the "amount of skin" visible in an
image. It is not based on the landmarker results, only on the
initial face detection, hence the fd (face detection) prefix.
landmarks:
Detected face landmarks.
roll_angle:
Roll angle, which indicates the amount of clockwise/anti-
clockwise rotation of the face relative to the image vertical
about the axis perpendicular to the face. Range [-180,180].
pan_angle:
Yaw angle, which indicates the leftward/rightward angle that
the face is pointing relative to the vertical plane
perpendicular to the image. Range [-180,180].
tilt_angle:
Pitch angle, which indicates the upwards/downwards angle that
the face is pointing relative to the image's horizontal plane.
Range [-180,180].
detection_confidence:
Detection confidence. Range [0, 1].
landmarking_confidence:
Face landmarking confidence. Range [0, 1].
joy_likelihood:
Joy likelihood.
sorrow_likelihood:
Sorrow likelihood.
anger_likelihood:
Anger likelihood.
surprise_likelihood:
Surprise likelihood.
under_exposed_likelihood:
Under-exposed likelihood.
blurred_likelihood:
Blurred likelihood.
headwear_likelihood:
Headwear likelihood.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.FaceAnnotation)
))
_sym_db.RegisterMessage(FaceAnnotation)
_sym_db.RegisterMessage(FaceAnnotation.Landmark)
LocationInfo = _reflection.GeneratedProtocolMessageType('LocationInfo', (_message.Message,), dict(
DESCRIPTOR = _LOCATIONINFO,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Detected entity location information.
Attributes:
lat_lng:
lat/long location coordinates.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.LocationInfo)
))
_sym_db.RegisterMessage(LocationInfo)
Property = _reflection.GeneratedProtocolMessageType('Property', (_message.Message,), dict(
DESCRIPTOR = _PROPERTY,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """A ``Property`` consists of a user-supplied name/value pair.
Attributes:
name:
Name of the property.
value:
Value of the property.
uint64_value:
Value of numeric properties.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.Property)
))
_sym_db.RegisterMessage(Property)
EntityAnnotation = _reflection.GeneratedProtocolMessageType('EntityAnnotation', (_message.Message,), dict(
DESCRIPTOR = _ENTITYANNOTATION,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Set of detected entity features.
Attributes:
mid:
Opaque entity ID. Some IDs may be available in `Google
Knowledge Graph Search API
<https://developers.google.com/knowledge-graph/>`__.
locale:
The language code for the locale in which the entity textual
``description`` is expressed.
description:
Entity textual description, expressed in its ``locale``
language.
score:
Overall score of the result. Range [0, 1].
confidence:
The accuracy of the entity detection in an image. For example,
for an image in which the "Eiffel Tower" entity is detected,
this field represents the confidence that there is a tower in
the query image. Range [0, 1].
topicality:
The relevancy of the ICA (Image Content Annotation) label to
the image. For example, the relevancy of "tower" is likely
higher to an image containing the detected "Eiffel Tower" than
to an image containing a detected distant towering building,
even though the confidence that there is a tower in each image
may be the same. Range [0, 1].
bounding_poly:
Image region to which this entity belongs. Not produced for
``LABEL_DETECTION`` features.
locations:
The location information for the detected entity. Multiple
``LocationInfo`` elements can be present because one location
may indicate the location of the scene in the image, and
another location may indicate the location of the place where
the image was taken. Location information is usually present
for landmarks.
properties:
Some entities may have optional user-supplied ``Property``
(name/value) fields, such a score or string that qualifies the
entity.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.EntityAnnotation)
))
_sym_db.RegisterMessage(EntityAnnotation)
SafeSearchAnnotation = _reflection.GeneratedProtocolMessageType('SafeSearchAnnotation', (_message.Message,), dict(
DESCRIPTOR = _SAFESEARCHANNOTATION,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Set of features pertaining to the image, computed by computer vision
methods over safe-search verticals (for example, adult, spoof, medical,
violence).
Attributes:
adult:
Represents the adult content likelihood for the image. Adult
content may contain elements such as nudity, pornographic
images or cartoons, or sexual activities.
spoof:
Spoof likelihood. The likelihood that an modification was made
to the image's canonical version to make it appear funny or
offensive.
medical:
Likelihood that this is a medical image.
violence:
Likelihood that this image contains violent content.
racy:
Likelihood that the request image contains racy content. Racy
content may include (but is not limited to) skimpy or sheer
clothing, strategically covered nudity, lewd or provocative
poses, or close-ups of sensitive body areas.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.SafeSearchAnnotation)
))
_sym_db.RegisterMessage(SafeSearchAnnotation)
LatLongRect = _reflection.GeneratedProtocolMessageType('LatLongRect', (_message.Message,), dict(
DESCRIPTOR = _LATLONGRECT,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Rectangle determined by min and max ``LatLng`` pairs.
Attributes:
min_lat_lng:
Min lat/long pair.
max_lat_lng:
Max lat/long pair.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.LatLongRect)
))
_sym_db.RegisterMessage(LatLongRect)
ColorInfo = _reflection.GeneratedProtocolMessageType('ColorInfo', (_message.Message,), dict(
DESCRIPTOR = _COLORINFO,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Color information consists of RGB channels, score, and the fraction of
the image that the color occupies in the image.
Attributes:
color:
RGB components of the color.
score:
Image-specific score for this color. Value in range [0, 1].
pixel_fraction:
The fraction of pixels the color occupies in the image. Value
in range [0, 1].
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.ColorInfo)
))
_sym_db.RegisterMessage(ColorInfo)
DominantColorsAnnotation = _reflection.GeneratedProtocolMessageType('DominantColorsAnnotation', (_message.Message,), dict(
DESCRIPTOR = _DOMINANTCOLORSANNOTATION,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Set of dominant colors and their corresponding scores.
Attributes:
colors:
RGB color values with their score and pixel fraction.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.DominantColorsAnnotation)
))
_sym_db.RegisterMessage(DominantColorsAnnotation)
ImageProperties = _reflection.GeneratedProtocolMessageType('ImageProperties', (_message.Message,), dict(
DESCRIPTOR = _IMAGEPROPERTIES,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Stores image properties, such as dominant colors.
Attributes:
dominant_colors:
If present, dominant colors completed successfully.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.ImageProperties)
))
_sym_db.RegisterMessage(ImageProperties)
CropHint = _reflection.GeneratedProtocolMessageType('CropHint', (_message.Message,), dict(
DESCRIPTOR = _CROPHINT,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Single crop hint that is used to generate a new crop when serving an
image.
Attributes:
bounding_poly:
The bounding polygon for the crop region. The coordinates of
the bounding box are in the original image's scale, as
returned in ``ImageParams``.
confidence:
Confidence of this being a salient region. Range [0, 1].
importance_fraction:
Fraction of importance of this salient region with respect to
the original image.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.CropHint)
))
_sym_db.RegisterMessage(CropHint)
CropHintsAnnotation = _reflection.GeneratedProtocolMessageType('CropHintsAnnotation', (_message.Message,), dict(
DESCRIPTOR = _CROPHINTSANNOTATION,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Set of crop hints that are used to generate new crops when serving
images.
Attributes:
crop_hints:
Crop hint results.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.CropHintsAnnotation)
))
_sym_db.RegisterMessage(CropHintsAnnotation)
CropHintsParams = _reflection.GeneratedProtocolMessageType('CropHintsParams', (_message.Message,), dict(
DESCRIPTOR = _CROPHINTSPARAMS,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Parameters for crop hints annotation request.
Attributes:
aspect_ratios:
Aspect ratios in floats, representing the ratio of the width
to the height of the image. For example, if the desired aspect
ratio is 4/3, the corresponding float value should be 1.33333.
If not specified, the best possible crop is returned. The
number of provided aspect ratios is limited to a maximum of
16; any aspect ratios provided after the 16th are ignored.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.CropHintsParams)
))
_sym_db.RegisterMessage(CropHintsParams)
WebDetectionParams = _reflection.GeneratedProtocolMessageType('WebDetectionParams', (_message.Message,), dict(
DESCRIPTOR = _WEBDETECTIONPARAMS,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Parameters for web detection request.
Attributes:
include_geo_results:
Whether to include results derived from the geo information in
the image.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.WebDetectionParams)
))
_sym_db.RegisterMessage(WebDetectionParams)
ImageContext = _reflection.GeneratedProtocolMessageType('ImageContext', (_message.Message,), dict(
DESCRIPTOR = _IMAGECONTEXT,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Image context and/or feature-specific parameters.
Attributes:
lat_long_rect:
lat/long rectangle that specifies the location of the image.
language_hints:
List of languages to use for TEXT\_DETECTION. In most cases,
an empty value yields the best results since it enables
automatic language detection. For languages based on the Latin
alphabet, setting ``language_hints`` is not needed. In rare
cases, when the language of the text in the image is known,
setting a hint will help get better results (although it will
be a significant hindrance if the hint is wrong). Text
detection returns an error if one or more of the specified
languages is not one of the `supported languages
</vision/docs/languages>`__.
crop_hints_params:
Parameters for crop hints annotation request.
web_detection_params:
Parameters for web detection.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.ImageContext)
))
_sym_db.RegisterMessage(ImageContext)
AnnotateImageRequest = _reflection.GeneratedProtocolMessageType('AnnotateImageRequest', (_message.Message,), dict(
DESCRIPTOR = _ANNOTATEIMAGEREQUEST,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Request for performing Google Cloud Vision API tasks over a
user-provided image, with user-requested features.
Attributes:
image:
The image to be processed.
features:
Requested features.
image_context:
Additional context that may accompany the image.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.AnnotateImageRequest)
))
_sym_db.RegisterMessage(AnnotateImageRequest)
AnnotateImageResponse = _reflection.GeneratedProtocolMessageType('AnnotateImageResponse', (_message.Message,), dict(
DESCRIPTOR = _ANNOTATEIMAGERESPONSE,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Response to an image annotation request.
Attributes:
face_annotations:
If present, face detection has completed successfully.
landmark_annotations:
If present, landmark detection has completed successfully.
logo_annotations:
If present, logo detection has completed successfully.
label_annotations:
If present, label detection has completed successfully.
text_annotations:
If present, text (OCR) detection has completed successfully.
full_text_annotation:
If present, text (OCR) detection or document (OCR) text
detection has completed successfully. This annotation provides
the structural hierarchy for the OCR detected text.
safe_search_annotation:
If present, safe-search annotation has completed successfully.
image_properties_annotation:
If present, image properties were extracted successfully.
crop_hints_annotation:
If present, crop hints have completed successfully.
web_detection:
If present, web detection has completed successfully.
error:
If set, represents the error message for the operation. Note
that filled-in image annotations are guaranteed to be correct,
even when ``error`` is set.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.AnnotateImageResponse)
))
_sym_db.RegisterMessage(AnnotateImageResponse)
BatchAnnotateImagesRequest = _reflection.GeneratedProtocolMessageType('BatchAnnotateImagesRequest', (_message.Message,), dict(
DESCRIPTOR = _BATCHANNOTATEIMAGESREQUEST,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Multiple image annotation requests are batched into a single service
call.
Attributes:
requests:
Individual image annotation requests for this batch.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.BatchAnnotateImagesRequest)
))
_sym_db.RegisterMessage(BatchAnnotateImagesRequest)
BatchAnnotateImagesResponse = _reflection.GeneratedProtocolMessageType('BatchAnnotateImagesResponse', (_message.Message,), dict(
DESCRIPTOR = _BATCHANNOTATEIMAGESRESPONSE,
__module__ = 'google.cloud.vision_v1p1beta1.proto.image_annotator_pb2'
,
__doc__ = """Response to a batch image annotation request.
Attributes:
responses:
Individual responses to image annotation requests within the
batch.
""",
# @@protoc_insertion_point(class_scope:google.cloud.vision.v1p1beta1.BatchAnnotateImagesResponse)
))
_sym_db.RegisterMessage(BatchAnnotateImagesResponse)
DESCRIPTOR.has_options = True
DESCRIPTOR._options = _descriptor._ParseOptions(descriptor_pb2.FileOptions(), _b('\n!com.google.cloud.vision.v1p1beta1B\023ImageAnnotatorProtoP\001ZCgoogle.golang.org/genproto/googleapis/cloud/vision/v1p1beta1;vision\370\001\001'))
try:
# THESE ELEMENTS WILL BE DEPRECATED.
# Please use the generated *_pb2_grpc.py files instead.
import grpc
from grpc.beta import implementations as beta_implementations
from grpc.beta import interfaces as beta_interfaces
from grpc.framework.common import cardinality
from grpc.framework.interfaces.face import utilities as face_utilities
class ImageAnnotatorStub(object):
"""Service that performs Google Cloud Vision API detection tasks over client
images, such as face, landmark, logo, label, and text detection. The
ImageAnnotator service returns detected entities from the images.
"""
def __init__(self, channel):
"""Constructor.
Args:
channel: A grpc.Channel.
"""
self.BatchAnnotateImages = channel.unary_unary(
'/google.cloud.vision.v1p1beta1.ImageAnnotator/BatchAnnotateImages',
request_serializer=BatchAnnotateImagesRequest.SerializeToString,
response_deserializer=BatchAnnotateImagesResponse.FromString,
)
class ImageAnnotatorServicer(object):
"""Service that performs Google Cloud Vision API detection tasks over client
images, such as face, landmark, logo, label, and text detection. The
ImageAnnotator service returns detected entities from the images.
"""
def BatchAnnotateImages(self, request, context):
"""Run image detection and annotation for a batch of images.
"""
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
context.set_details('Method not implemented!')
raise NotImplementedError('Method not implemented!')
def add_ImageAnnotatorServicer_to_server(servicer, server):
rpc_method_handlers = {
'BatchAnnotateImages': grpc.unary_unary_rpc_method_handler(
servicer.BatchAnnotateImages,
request_deserializer=BatchAnnotateImagesRequest.FromString,
response_serializer=BatchAnnotateImagesResponse.SerializeToString,
),
}
generic_handler = grpc.method_handlers_generic_handler(
'google.cloud.vision.v1p1beta1.ImageAnnotator', rpc_method_handlers)
server.add_generic_rpc_handlers((generic_handler,))
class BetaImageAnnotatorServicer(object):
"""The Beta API is deprecated for 0.15.0 and later.
It is recommended to use the GA API (classes and functions in this
file not marked beta) for all further purposes. This class was generated
only to ease transition from grpcio<0.15.0 to grpcio>=0.15.0."""
"""Service that performs Google Cloud Vision API detection tasks over client
images, such as face, landmark, logo, label, and text detection. The
ImageAnnotator service returns detected entities from the images.
"""
def BatchAnnotateImages(self, request, context):
"""Run image detection and annotation for a batch of images.
"""
context.code(beta_interfaces.StatusCode.UNIMPLEMENTED)
class BetaImageAnnotatorStub(object):
"""The Beta API is deprecated for 0.15.0 and later.
It is recommended to use the GA API (classes and functions in this
file not marked beta) for all further purposes. This class was generated
only to ease transition from grpcio<0.15.0 to grpcio>=0.15.0."""
"""Service that performs Google Cloud Vision API detection tasks over client
images, such as face, landmark, logo, label, and text detection. The
ImageAnnotator service returns detected entities from the images.
"""
def BatchAnnotateImages(self, request, timeout, metadata=None, with_call=False, protocol_options=None):
"""Run image detection and annotation for a batch of images.
"""
raise NotImplementedError()
BatchAnnotateImages.future = None
def beta_create_ImageAnnotator_server(servicer, pool=None, pool_size=None, default_timeout=None, maximum_timeout=None):
"""The Beta API is deprecated for 0.15.0 and later.
It is recommended to use the GA API (classes and functions in this
file not marked beta) for all further purposes. This function was
generated only to ease transition from grpcio<0.15.0 to grpcio>=0.15.0"""
request_deserializers = {
('google.cloud.vision.v1p1beta1.ImageAnnotator', 'BatchAnnotateImages'): BatchAnnotateImagesRequest.FromString,
}
response_serializers = {
('google.cloud.vision.v1p1beta1.ImageAnnotator', 'BatchAnnotateImages'): BatchAnnotateImagesResponse.SerializeToString,
}
method_implementations = {
('google.cloud.vision.v1p1beta1.ImageAnnotator', 'BatchAnnotateImages'): face_utilities.unary_unary_inline(servicer.BatchAnnotateImages),
}
server_options = beta_implementations.server_options(request_deserializers=request_deserializers, response_serializers=response_serializers, thread_pool=pool, thread_pool_size=pool_size, default_timeout=default_timeout, maximum_timeout=maximum_timeout)
return beta_implementations.server(method_implementations, options=server_options)
def beta_create_ImageAnnotator_stub(channel, host=None, metadata_transformer=None, pool=None, pool_size=None):
"""The Beta API is deprecated for 0.15.0 and later.
It is recommended to use the GA API (classes and functions in this
file not marked beta) for all further purposes. This function was
generated only to ease transition from grpcio<0.15.0 to grpcio>=0.15.0"""
request_serializers = {
('google.cloud.vision.v1p1beta1.ImageAnnotator', 'BatchAnnotateImages'): BatchAnnotateImagesRequest.SerializeToString,
}
response_deserializers = {
('google.cloud.vision.v1p1beta1.ImageAnnotator', 'BatchAnnotateImages'): BatchAnnotateImagesResponse.FromString,
}
cardinalities = {
'BatchAnnotateImages': cardinality.Cardinality.UNARY_UNARY,
}
stub_options = beta_implementations.stub_options(host=host, metadata_transformer=metadata_transformer, request_serializers=request_serializers, response_deserializers=response_deserializers, thread_pool=pool, thread_pool_size=pool_size)
return beta_implementations.dynamic_stub(channel, 'google.cloud.vision.v1p1beta1.ImageAnnotator', cardinalities, options=stub_options)
except ImportError:
pass
# @@protoc_insertion_point(module_scope)
| 42.973206 | 9,043 | 0.743392 | [
"Apache-2.0"
] | Alexander-Minyushkin/google-cloud-python | vision/google/cloud/vision_v1p1beta1/proto/image_annotator_pb2.py | 89,814 | Python |
import logging
# noinspection PyPackageRequirements
from telegram.ext import CommandHandler, ConversationHandler
# noinspection PyPackageRequirements
from telegram import ChatAction, Update
from bot import stickersbot
from bot.utils import decorators
from bot.utils import utils
from bot.database.base import session_scope
from bot.database.models.pack import Pack
from bot.strings import Strings
logger = logging.getLogger(__name__)
@decorators.action(ChatAction.TYPING)
@decorators.restricted
@decorators.failwithmessage
def on_list_command(update: Update, _):
logger.info('/list')
# packs = db.get_user_packs(update.effective_user.id, as_namedtuple=True)
with session_scope() as session:
packs = session.query(Pack).filter_by(user_id=update.effective_user.id).order_by(Pack.title).all()
packs = packs[:98] # can't include more than 100 entities
strings_list = ['<a href="{}">{}</a> ({})'.format(utils.name2link(pack.name), pack.title, 'a' if pack.is_animated else 's') for pack in packs]
if not strings_list:
update.message.reply_text(Strings.LIST_NO_PACKS)
return
update.message.reply_html('• {}'.format('\n• '.join(strings_list)) + Strings.LIST_FOOTER)
return ConversationHandler.END # /list should end whatever conversation the user was having
stickersbot.add_handler(CommandHandler(['list', 'l'], on_list_command))
| 34.975 | 150 | 0.754825 | [
"MIT"
] | Ankit29-A/sticker-thief | bot/handlers/packs/list.py | 1,403 | Python |
# -*- coding: utf-8 -*-
'''
QSDsan: Quantitative Sustainable Design for sanitation and resource recovery systems
This module is developed by:
Joy Zhang <[email protected]>
This module is under the University of Illinois/NCSA Open Source License.
Please refer to https://github.com/QSD-Group/QSDsan/blob/main/LICENSE.txt
for license details.
'''
from .. import SanUnit, WasteStream, Process, Processes, CompiledProcesses
from ._clarifier import _settling_flux
from sympy import symbols, lambdify, Matrix
from scipy.integrate import solve_ivp
from warnings import warn
from math import floor, ceil
import numpy as np
import pandas as pd
from numba import njit
__all__ = ('CSTR',
'SBR',
# 'PFR',
)
def _add_aeration_to_growth_model(aer, model):
if isinstance(aer, Process):
processes = Processes(model.tuple)
processes.append(aer)
processes.compile()
else:
processes = model
processes.compile()
return processes
# %%
@njit(cache=True)
def dydt_cstr_no_rxn_fixed_aer(QC_ins, dQC_ins, V_arr, Q_e_arr, _dstate, Cs):
Q_ins = QC_ins[:, -1]
C_ins = QC_ins[:, :-1]
flow_in = Q_ins @ C_ins / V_arr
Q_e_arr[:] = Q_ins.sum(axis=0)
_dstate[-1] = dQC_ins[:, -1].sum(axis=0)
flow_out = Q_e_arr * Cs / V_arr
_dstate[:-1] = flow_in - flow_out
@njit(cache=True)
def dydt_cstr_no_rxn_controlled_aer(QC_ins, dQC_ins, V_arr, Q_e_arr, _dstate, Cs):
Q_ins = QC_ins[:, -1]
C_ins = QC_ins[:, :-1]
flow_in = Q_ins @ C_ins / V_arr
Q_e_arr[:] = Q_ins.sum(axis=0)
_dstate[-1] = dQC_ins[:, -1].sum(axis=0)
flow_out = Q_e_arr * Cs / V_arr
_dstate[:-1] = flow_in - flow_out
#%%
class CSTR(SanUnit):
'''
A single continuous stirred tank reactor.
Parameters
----------
ID : str
ID for the reactor.
ins : :class:`WasteStream`
Influents to the reactor. Can be an array of up to 3 WasteStream objects by
default, typically wastewater to be treated, recycled effluent, recycled
activated sludge.
outs : :class:`WasteStream`
Treated effluent.
split : iterable of float
Volumetric splits of effluent flows if there are more than one effluent.
The default is None.
V_max : float
Designed volume, in [m^3]. The default is 1000.
aeration : float or :class:`Process`, optional
Aeration setting. Either specify a targeted dissolved oxygen concentration
in [mg O2/L] or provide a :class:`Process` object to represent aeration,
or None for no aeration. The default is 2.0.
DO_ID : str, optional
The :class:`Component` ID for dissolved oxygen, only relevant when the
reactor is aerated. The default is 'S_O2'.
suspended_growth_model : :class:`Processes`, optional
The suspended growth biokinetic model. The default is None.
'''
_N_ins = 3
_N_outs = 1
_ins_size_is_fixed = False
_outs_size_is_fixed = False
def __init__(self, ID='', ins=None, outs=(), split=None, thermo=None,
init_with='WasteStream', V_max=1000, aeration=2.0,
DO_ID='S_O2', suspended_growth_model=None,
isdynamic=True, **kwargs):
SanUnit.__init__(self, ID, ins, outs, thermo, init_with, isdynamic=isdynamic)
self._V_max = V_max
self._aeration = aeration
self._DO_ID = DO_ID
self._model = suspended_growth_model
self._concs = None
self._mixed = WasteStream()
self.split = split
for attr, value in kwargs.items():
setattr(self, attr, value)
@property
def V_max(self):
'''[float] The designed maximum liquid volume, not accounting for increased volume due to aeration, in m^3.'''
return self._V_max
@V_max.setter
def V_max(self, Vm):
self._V_max = Vm
@property
def aeration(self):
'''[:class:`Process` or float or NoneType] Aeration model.'''
return self._aeration
@aeration.setter
def aeration(self, ae):
if ae is None or isinstance(ae, Process): self._aeration = ae
elif isinstance(ae, (float, int)):
if ae < 0:
raise ValueError('targeted dissolved oxygen concentration for aeration must be non-negative.')
else:
if ae > 14:
warn(f'targeted dissolved oxygen concentration for {self.ID} might exceed the saturated level.')
self._aeration = ae
else:
raise TypeError(f'aeration must be one of the following types: float, '
f'int, Process, NoneType. Not {type(ae)}')
@property
def suspended_growth_model(self):
'''[:class:`CompiledProcesses` or NoneType] Suspended growth model.'''
return self._model
@suspended_growth_model.setter
def suspended_growth_model(self, model):
if isinstance(model, CompiledProcesses) or model is None: self._model = model
else: raise TypeError(f'suspended_growth_model must be one of the following '
f'types: CompiledProesses, NoneType. Not {type(model)}')
@property
def DO_ID(self):
'''[str] The `Component` ID for dissolved oxygen used in the suspended growth model and the aeration model.'''
return self._DO_ID
@DO_ID.setter
def DO_ID(self, doid):
if doid not in self.components.IDs:
raise ValueError(f'DO_ID must be in the set of `CompiledComponents` used to set thermo, '
f'i.e., one of {self.components.IDs}.')
self._DO_ID = doid
@property
def split(self):
'''[numpy.1darray or NoneType] The volumetric split of outs.'''
return self._split
@split.setter
def split(self, split):
if split is None: self._split = split
else:
if len(split) != len(self._outs):
raise ValueError('split and outs must have the same size')
self._split = np.array(split)/sum(split)
@property
def state(self):
'''The state of the CSTR, including component concentrations [mg/L] and flow rate [m^3/d].'''
if self._state is None: return None
else:
return dict(zip(list(self.components.IDs) + ['Q'], self._state))
@state.setter
def state(self, QCs):
QCs = np.asarray(QCs)
if QCs.shape != (len(self.components)+1, ):
raise ValueError(f'state must be a 1D array of length {len(self.components) + 1},'
'indicating component concentrations [mg/L] and total flow rate [m^3/d]')
self._state = QCs
def set_init_conc(self, **kwargs):
'''set the initial concentrations [mg/L] of the CSTR.'''
Cs = np.zeros(len(self.components))
cmpx = self.components.index
for k, v in kwargs.items(): Cs[cmpx(k)] = v
self._concs = Cs
def _init_state(self):
mixed = self._mixed
Q = mixed.get_total_flow('m3/d')
if self._concs is not None: Cs = self._concs
else: Cs = mixed.conc
self._state = np.append(Cs, Q).astype('float64')
self._dstate = self._state * 0.
def _update_state(self):
arr = self._state
if self.split is None: self._outs[0].state = arr
else:
for ws, spl in zip(self._outs, self.split):
y = arr.copy()
y[-1] *= spl
ws.state = y
def _update_dstate(self):
arr = self._dstate
if self.split is None: self._outs[0].dstate = arr
else:
for ws, spl in zip(self._outs, self.split):
y = arr.copy()
y[-1] *= spl
ws.dstate = y
def _run(self):
'''Only to converge volumetric flows.'''
mixed = self._mixed # avoid creating multiple new streams
mixed.mix_from(self.ins)
Q = mixed.F_vol # m3/hr
if self.split is None: self.outs[0].copy_like(mixed)
else:
for ws, spl in zip(self._outs, self.split):
ws.copy_like(mixed)
ws.set_total_flow(Q*spl, 'm3/hr')
def get_retained_mass(self, biomass_IDs):
cmps = self.components
mass = cmps.i_mass * self._state[:-1]
return self._V_max * mass[cmps.indices(biomass_IDs)].sum()
@property
def ODE(self):
if self._ODE is None:
self._compile_ODE()
return self._ODE
def _compile_ODE(self):
isa = isinstance
C = list(symbols(self.components.IDs))
m = len(C)
if self._model is None:
warn(f'{self.ID} was initialized without a suspended growth model, '
f'and thus run as a non-reactive unit')
r = lambda *args: np.zeros(m)
else:
processes = _add_aeration_to_growth_model(self._aeration, self._model)
r_eqs = list(processes.production_rates.rate_of_production)
r = lambdify(C, r_eqs, 'numpy')
_dstate = self._dstate
_update_dstate = self._update_dstate
V_arr = np.full(m, self._V_max)
Q_e_arr = np.zeros(m)
if isa(self._aeration, (float, int)):
i = self.components.index(self._DO_ID)
fixed_DO = self._aeration
def dy_dt(t, QC_ins, QC, dQC_ins):
Cs = QC[:-1]
Cs[i] = fixed_DO
dydt_cstr_no_rxn_controlled_aer(QC_ins, dQC_ins, V_arr, Q_e_arr, _dstate, Cs)
_dstate[:-1] += r(*Cs)
_dstate[i] = 0
_update_dstate()
else:
def dy_dt(t, QC_ins, QC, dQC_ins):
Cs = QC[:-1]
dydt_cstr_no_rxn_fixed_aer(QC_ins, dQC_ins, V_arr, Q_e_arr, _dstate, Cs)
_dstate[:-1] += r(*Cs)
_update_dstate()
self._ODE = dy_dt
def _design(self):
pass
class SBR(SanUnit):
'''
Sequential batch reactors operated in parallel. The number of reactors is
determined by operation cycle and influent flowrate. [1]_
Parameters
----------
ID : str
ID for the reactors. The default is ''.
ins : :class:`WasteStream`
Influent to the reactor. Expected number of influent is 1.
outs : :class:`WasteStream`
Treated effluent and wasted sludge.
surface_area : float, optional
Surface area of the reactor bottom, in [m^2]. The reactor is assumed
to be cylinder. The default is 1500.
height : float, optional
Height of the reactor, in [m]. The default is 4.
operation_cycle : iterable of float, optional
Operation cycle of the SBR, time for each stage specified in [h]. There
are 7 stages: 1 - fill, 2 - fill, 3 - mix, 4 - mix, 5 - settle, 6 - decant,
7 - desludge. The first 4 stages are modeled as a biological reactor.
The 5th stage is modeled as a 1D N-layer settler. The last 2 stages are
assumed inactive. The default is (0.5, 1.5, 2.0, 0, 1.0, 0.5, 0.1).
aeration : iterable of float and/or :class:`Process`, optional
Aeration settings for the first 4 stages of the operation cycle. Either
specify a targeted dissolved oxygen concentration in [mg O2/L] or provide
a :class:`Process` object to represent aeration, or None for no aeration.
The default is (None, None, None, 2.0).
DO_ID : str, optional
The :class:`Component` ID for dissolved oxygen, only relevant when the
reactor is aerated. The default is 'S_O2'.
suspended_growth_model : :class:`Processes`, optional
The suspended growth biokinetic model. The default is None.
N_layer : int, optional
The number of layers to model settling. The default is 10.
pumped_flow : float, optional
Designed effluent flowrate, in [m^3/d]. The default is None.
underflow : float, optional
Designed wasted activated sludge flowrate, in [m^3/d]. The default is None.
X_threshold : float, optional
Threshold suspended solid concentration, in [g/m^3]. The default is 3000.
v_max : float, optional
Maximum theoretical (i.e. Vesilind) settling velocity, in [m/d]. The
default is 474.
v_max_practical : float, optional
Maximum practical settling velocity, in [m/d]. The default is 250.
rh : float, optional
Hindered zone settling parameter in the double-exponential settling velocity
function, in [m^3/g]. The default is 5.76e-4.
rp : float, optional
Flocculant zone settling parameter in the double-exponential settling velocity
function, in [m^3/g]. The default is 2.86e-3.
fns : float, optional
Non-settleable fraction of the suspended solids, dimensionless. Must be within
[0, 1]. The default is 2.28e-3.
cache_state : bool, optional
Whether to store volume and composition of retained sludge in the tank from
most recent run. The default is True.
References
----------
.. [1] Takács, I.; Patry, G. G.; Nolasco, D. A Dynamic Model of the Clarification
-Thickening Process. Water Res. 1991, 25 (10), 1263–1271.
https://doi.org/10.1016/0043-1354(91)90066-Y.
'''
_N_ins = 1
_N_outs = 2
def __init__(self, ID='', ins=None, outs=(), thermo=None, init_with='WasteStream',
surface_area=1500, height=4,
operation_cycle=(0.5, 1.5, 2.0, 0, 1.0, 0.5, 0.1),
aeration=(None, None, None, 2.0), DO_ID='S_O2',
suspended_growth_model=None, N_layer=10,
pumped_flow=None, underflow=None,
X_threshold=3000, v_max=474, v_max_practical=250,
rh=5.76e-4, rp=2.86e-3, fns=2.28e-3,
cache_state=True, **kwargs):
SanUnit.__init__(self, ID, ins, outs, thermo, init_with)
self._V = surface_area * height
self._A = surface_area
self._h = height
self._operation_cycle = operation_cycle
self._aeration = aeration
self._DO_ID = DO_ID
self._model = suspended_growth_model
self._N_layer = N_layer
self._Q_e = pumped_flow
self._Q_WAS = underflow
self._X_t = X_threshold
self._v_max = v_max
self._v_max_p = v_max_practical
self._rh = rh
self._rp = rp
self._fns = fns
self._cache_state = cache_state
for attr, value in kwargs.items():
setattr(self, attr, value)
self._init_Vas = None
self._init_Cas = None
self._dynamic_composition = None
@property
def operation_cycle(self):
return dict(zip(('fill_1', 'fill_2', 'mix_1', 'mix_2', 'settle', 'decant', 'desludge'),
self._operation_cycle))
@property
def total_cycle_time(self):
return sum(self._operation_cycle)
@property
def aeration(self):
return dict(zip(('fill_1', 'fill_2', 'mix_1', 'mix_2'),
self._aeration[:4]))
@property
def C_t(self):
if self._dynamic_composition:
return pd.DataFrame(self._dynamic_composition,
columns = ['Time[d]'] + list(self.components.IDs))
else: return None
def _run(self, cache_state=True):
if self._model is None:
raise RuntimeError(f'{self.ID} was initialized without a suspended growth model.')
else:
isa = isinstance
inf = self.ins[0]
Q_in = inf.get_total_flow('m3/d')
eff, sludge = self.outs
eff.copy_like(inf)
sludge.copy_like(inf)
C_in = inf.mass / inf.F_vol * 1e3 # concentrations in g/m3
cmps = self.components
C = list(symbols(cmps.IDs))
if self._init_Vas is not None:
V_0 = self._init_Vas
C_0 = self._init_Cas
else:
V_0 = 0
C_0 = C_in
n = self._N_layer
if self._aeration.count(None) == len(self._aeration):
Vmax = self._V
hj = self._h/n
else:
Vmax = self._V*0.75
hj = self._h*0.75/n
# ********fill and mix/aerate stages***********
T_fill = (Vmax - V_0)/Q_in # maximum total fill time in day
T = [t/24 for t in self._operation_cycle] # operation cycle in day
if T_fill <= T[0]:
schedule = [T_fill, T[0]-T_fill] + T[1:4]
aer = [self._aeration[0], self._aeration[0]] + list(self._aeration[1:4])
fill = [True] + [False]*4
V_total = Vmax
elif T_fill <= T[0]+T[1]:
schedule = [T[0], T_fill-T[0], T[0]+T[1]-T_fill] + T[2:4]
aer = list(self._aeration[:2]) + [self._aeration[1]] + list(self._aeration[2:4])
fill = [True]*2 + [False]*3
V_total = Vmax
else:
schedule = T[:4]
aer = list(self._aeration[:4])
fill = [True]*2 + [False]*2
V_total = Q_in*(T[0]+T[1])+V_0
hj = V_total/self._V*self._h/n
for i in range(1, len(schedule)):
if fill[-i] == fill[-i-1] and aer[-i] == aer[-i-1]:
schedule[-i-1] += schedule[-i]
schedule[-i] = 0
t_arr = np.array([])
y_mat = np.ndarray([])
for i in range(len(schedule)):
if schedule[i] > 0:
dC_dt, J_func = self._compile_dC_dt(V_0, Q_in, C_in, C, fill[i], aer[i])
if isa(aer[i], (float, int)): C_0[cmps.index(self._DO_ID)] = aer[i]
sol = solve_ivp(dC_dt, (0, schedule[i]), C_0, method='BDF', jac=J_func)
C_0 = sol.y.transpose()[-1]
V_0 += Q_in * schedule[i] * fill[i]
t_arr = np.concatenate((t_arr, sol.t + t_arr[-1]))
y_mat = np.hstack((y_mat, sol.y))
self._dynamic_composition = np.vstack((t_arr, y_mat)).transpose()
# *********settle, decant, desludge**********
eff.set_flow(C_0*eff.F_vol, 'g/hr', self.components.IDs)
X_0 = eff.get_TSS()
X_min = X_0 * self._fns
T_settle = T[4]
def dX_dt(t, X):
VX = [_settling_flux(x, self._v_max, self._v_max_p, X_min, self._rh, self._rp) for x in X]
J = [VX[j] if X[j+1] <= self._X_t else min(VX[j], VX[j+1]) for j in range(n-1)]
settle_out = np.array(J + [0])
settle_in = np.array([0] + J)
dXdt = (settle_in - settle_out)/hj
return dXdt
sol = solve_ivp(dX_dt, (0, T_settle), np.ones(n)*X_0)
X = sol.y.transpose()[-1]
V_eff = min(T[5]*self._Q_e, V_total*(n-1)/n)
n_eff = V_eff/V_total
w_eff = np.array([1]*floor(n_eff)+[n_eff-floor(n_eff)])
X_eff = np.average(X[:ceil(n_eff)], weights=w_eff)
eff_mass_flow = (X_eff/X_0*cmps.x + (1-cmps.x))*C_0*V_eff/T[5]
eff.set_flow(eff_mass_flow, 'g/d', cmps.IDs)
V_was = min(T[6]*self._Q_WAS, V_total-V_eff)
X_as = (V_total*X_0 - V_eff*X_eff) / (V_total-V_eff)
C_as = (X_as/X_0*cmps.x + (1-cmps.x))*C_0
was_mass_flow = C_as*V_was/T[6]
sludge.set_flow(was_mass_flow, 'g/d', cmps.IDs)
if self._cache_state:
self._init_Vas = V_total - V_eff - V_was
self._init_Cas = C_as
def _design(self):
pass
def _compile_dC_dt(self, V0, Qin, Cin, C, fill, aer):
isa = isinstance
processes = _add_aeration_to_growth_model(aer, self._model)
if fill:
t = symbols('t')
mass_balance_terms = list(zip(Cin, C, processes.production_rates.rate_of_production))
C_dot_eqs = [(cin-c)/(t+V0/Qin) + r for cin, c, r in mass_balance_terms]
if isa(aer, (float, int)): C_dot_eqs[self.components.index(self._DO_ID)] = 0
def dC_dt(t, y):
C_dot = lambdify([t]+C, C_dot_eqs)
return C_dot(t, *y)
J = Matrix(dC_dt(t, C)).jacobian(C)
else:
C_dot_eqs = processes.production_rates.rate_of_production
if isa(aer, (float, int)): C_dot_eqs[self.components.index(self._DO_ID)] = 0
def dC_dt(t, y):
C_dot = lambdify(C, C_dot_eqs)
return C_dot(*y)
J = Matrix(dC_dt(None, C)).jacobian(C)
def J_func(t, y):
J_func = lambdify(C, J)
return J_func(*y)
return (dC_dt, J_func)
# class PFR(SanUnit):
# _N_ins = 1
# _N_outs = 2
# def __init__(self, ID='', ins=None, outs=(), **kwargs):
# SanUnit.__init__(self, ID, ins, outs)
# def _run(self, steady_state=True):
# pass
# def _design(self):
# pass | 38.412727 | 118 | 0.576229 | [
"Unlicense"
] | QSD-for-WaSH/sanitation | qsdsan/sanunits/_suspended_growth_bioreactor.py | 21,130 | Python |
# vim: set et sw=4 sts=4 fileencoding=utf-8:
#
# Python camera library for the Rasperry-Pi camera module
# Copyright (c) 2013-2017 Dave Jones <[email protected]>
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# * Neither the name of the copyright holder nor the
# names of its contributors may be used to endorse or promote products
# derived from this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
# POSSIBILITY OF SUCH DAMAGE.
from __future__ import (
unicode_literals,
print_function,
division,
absolute_import,
)
# Make Py2's str equivalent to Py3's
str = type('')
import warnings
import datetime
import mimetypes
import ctypes as ct
import threading
from fractions import Fraction
from operator import itemgetter
from collections import namedtuple
from . import bcm_host, mmal, mmalobj as mo
from .exc import (
PiCameraError,
PiCameraValueError,
PiCameraRuntimeError,
PiCameraClosed,
PiCameraNotRecording,
PiCameraAlreadyRecording,
PiCameraMMALError,
PiCameraDeprecated,
PiCameraFallback,
)
from .encoders import (
PiVideoFrame,
PiVideoEncoder,
PiRawVideoEncoder,
PiCookedVideoEncoder,
PiRawOneImageEncoder,
PiRawMultiImageEncoder,
PiCookedOneImageEncoder,
PiCookedMultiImageEncoder,
)
from .renderers import (
PiPreviewRenderer,
PiOverlayRenderer,
PiNullSink,
)
from .color import Color
try:
from RPi import GPIO
except ImportError:
# Can't find RPi.GPIO so just null-out the reference
GPIO = None
def docstring_values(values, indent=8):
"""
Formats a dictionary of values for inclusion in a docstring.
"""
return ('\n' + ' ' * indent).join(
"* ``'%s'``" % k
for (k, v) in
sorted(values.items(), key=itemgetter(1)))
class PiCameraMaxResolution(object):
"""
Singleton representing the maximum resolution of the camera module.
"""
PiCameraMaxResolution = PiCameraMaxResolution()
class PiCameraMaxFramerate(object):
"""
Singleton representing the maximum framerate of the camera module.
"""
PiCameraMaxFramerate = PiCameraMaxFramerate()
class PiCamera(object):
"""
Provides a pure Python interface to the Raspberry Pi's camera module.
Upon construction, this class initializes the camera. The *camera_num*
parameter (which defaults to 0) selects the camera module that the instance
will represent. Only the Raspberry Pi compute module currently supports
more than one camera.
The *sensor_mode*, *resolution*, *framerate*, *framerate_range*, and
*clock_mode* parameters provide initial values for the :attr:`sensor_mode`,
:attr:`resolution`, :attr:`framerate`, :attr:`framerate_range`, and
:attr:`clock_mode` attributes of the class (these attributes are all
relatively expensive to set individually, hence setting them all upon
construction is a speed optimization). Please refer to the attribute
documentation for more information and default values.
The *stereo_mode* and *stereo_decimate* parameters configure dual cameras
on a compute module for sterescopic mode. These parameters can only be set
at construction time; they cannot be altered later without closing the
:class:`PiCamera` instance and recreating it. The *stereo_mode* parameter
defaults to ``'none'`` (no stereoscopic mode) but can be set to
``'side-by-side'`` or ``'top-bottom'`` to activate a stereoscopic mode. If
the *stereo_decimate* parameter is ``True``, the resolution of the two
cameras will be halved so that the resulting image has the same dimensions
as if stereoscopic mode were not being used.
The *led_pin* parameter can be used to specify the GPIO pin which should be
used to control the camera's LED via the :attr:`led` attribute. If this is
not specified, it should default to the correct value for your Pi platform.
You should only need to specify this parameter if you are using a custom
DeviceTree blob (this is only typical on the `Compute Module`_ platform).
No preview or recording is started automatically upon construction. Use
the :meth:`capture` method to capture images, the :meth:`start_recording`
method to begin recording video, or the :meth:`start_preview` method to
start live display of the camera's input.
Several attributes are provided to adjust the camera's configuration. Some
of these can be adjusted while a recording is running, like
:attr:`brightness`. Others, like :attr:`resolution`, can only be adjusted
when the camera is idle.
When you are finished with the camera, you should ensure you call the
:meth:`close` method to release the camera resources::
camera = PiCamera()
try:
# do something with the camera
pass
finally:
camera.close()
The class supports the context manager protocol to make this particularly
easy (upon exiting the :keyword:`with` statement, the :meth:`close` method
is automatically called)::
with PiCamera() as camera:
# do something with the camera
pass
.. versionchanged:: 1.8
Added *stereo_mode* and *stereo_decimate* parameters.
.. versionchanged:: 1.9
Added *resolution*, *framerate*, and *sensor_mode* parameters.
.. versionchanged:: 1.10
Added *led_pin* parameter.
.. versionchanged:: 1.11
Added *clock_mode* parameter, and permitted setting of resolution as
appropriately formatted string.
.. versionchanged:: 1.13
Added *framerate_range* parameter.
.. _Compute Module: https://www.raspberrypi.org/documentation/hardware/computemodule/cmio-camera.md
"""
CAMERA_PREVIEW_PORT = 0
CAMERA_VIDEO_PORT = 1
CAMERA_CAPTURE_PORT = 2
MAX_RESOLUTION = PiCameraMaxResolution # modified by PiCamera.__init__
MAX_FRAMERATE = PiCameraMaxFramerate # modified by PiCamera.__init__
DEFAULT_ANNOTATE_SIZE = 32
CAPTURE_TIMEOUT = 60
METER_MODES = {
'average': mmal.MMAL_PARAM_EXPOSUREMETERINGMODE_AVERAGE,
'spot': mmal.MMAL_PARAM_EXPOSUREMETERINGMODE_SPOT,
'backlit': mmal.MMAL_PARAM_EXPOSUREMETERINGMODE_BACKLIT,
'matrix': mmal.MMAL_PARAM_EXPOSUREMETERINGMODE_MATRIX,
}
EXPOSURE_MODES = {
'off': mmal.MMAL_PARAM_EXPOSUREMODE_OFF,
'auto': mmal.MMAL_PARAM_EXPOSUREMODE_AUTO,
'night': mmal.MMAL_PARAM_EXPOSUREMODE_NIGHT,
'nightpreview': mmal.MMAL_PARAM_EXPOSUREMODE_NIGHTPREVIEW,
'backlight': mmal.MMAL_PARAM_EXPOSUREMODE_BACKLIGHT,
'spotlight': mmal.MMAL_PARAM_EXPOSUREMODE_SPOTLIGHT,
'sports': mmal.MMAL_PARAM_EXPOSUREMODE_SPORTS,
'snow': mmal.MMAL_PARAM_EXPOSUREMODE_SNOW,
'beach': mmal.MMAL_PARAM_EXPOSUREMODE_BEACH,
'verylong': mmal.MMAL_PARAM_EXPOSUREMODE_VERYLONG,
'fixedfps': mmal.MMAL_PARAM_EXPOSUREMODE_FIXEDFPS,
'antishake': mmal.MMAL_PARAM_EXPOSUREMODE_ANTISHAKE,
'fireworks': mmal.MMAL_PARAM_EXPOSUREMODE_FIREWORKS,
}
FLASH_MODES = {
'off': mmal.MMAL_PARAM_FLASH_OFF,
'auto': mmal.MMAL_PARAM_FLASH_AUTO,
'on': mmal.MMAL_PARAM_FLASH_ON,
'redeye': mmal.MMAL_PARAM_FLASH_REDEYE,
'fillin': mmal.MMAL_PARAM_FLASH_FILLIN,
'torch': mmal.MMAL_PARAM_FLASH_TORCH,
}
AWB_MODES = {
'off': mmal.MMAL_PARAM_AWBMODE_OFF,
'auto': mmal.MMAL_PARAM_AWBMODE_AUTO,
'sunlight': mmal.MMAL_PARAM_AWBMODE_SUNLIGHT,
'cloudy': mmal.MMAL_PARAM_AWBMODE_CLOUDY,
'shade': mmal.MMAL_PARAM_AWBMODE_SHADE,
'tungsten': mmal.MMAL_PARAM_AWBMODE_TUNGSTEN,
'fluorescent': mmal.MMAL_PARAM_AWBMODE_FLUORESCENT,
'incandescent': mmal.MMAL_PARAM_AWBMODE_INCANDESCENT,
'flash': mmal.MMAL_PARAM_AWBMODE_FLASH,
'horizon': mmal.MMAL_PARAM_AWBMODE_HORIZON,
}
IMAGE_EFFECTS = {
'none': mmal.MMAL_PARAM_IMAGEFX_NONE,
'negative': mmal.MMAL_PARAM_IMAGEFX_NEGATIVE,
'solarize': mmal.MMAL_PARAM_IMAGEFX_SOLARIZE,
# The following don't work
#'posterize': mmal.MMAL_PARAM_IMAGEFX_POSTERIZE,
#'whiteboard': mmal.MMAL_PARAM_IMAGEFX_WHITEBOARD,
#'blackboard': mmal.MMAL_PARAM_IMAGEFX_BLACKBOARD,
'sketch': mmal.MMAL_PARAM_IMAGEFX_SKETCH,
'denoise': mmal.MMAL_PARAM_IMAGEFX_DENOISE,
'emboss': mmal.MMAL_PARAM_IMAGEFX_EMBOSS,
'oilpaint': mmal.MMAL_PARAM_IMAGEFX_OILPAINT,
'hatch': mmal.MMAL_PARAM_IMAGEFX_HATCH,
'gpen': mmal.MMAL_PARAM_IMAGEFX_GPEN,
'pastel': mmal.MMAL_PARAM_IMAGEFX_PASTEL,
'watercolor': mmal.MMAL_PARAM_IMAGEFX_WATERCOLOUR,
'film': mmal.MMAL_PARAM_IMAGEFX_FILM,
'blur': mmal.MMAL_PARAM_IMAGEFX_BLUR,
'saturation': mmal.MMAL_PARAM_IMAGEFX_SATURATION,
'colorswap': mmal.MMAL_PARAM_IMAGEFX_COLOURSWAP,
'washedout': mmal.MMAL_PARAM_IMAGEFX_WASHEDOUT,
'posterise': mmal.MMAL_PARAM_IMAGEFX_POSTERISE,
'colorpoint': mmal.MMAL_PARAM_IMAGEFX_COLOURPOINT,
'colorbalance': mmal.MMAL_PARAM_IMAGEFX_COLOURBALANCE,
'cartoon': mmal.MMAL_PARAM_IMAGEFX_CARTOON,
'deinterlace1': mmal.MMAL_PARAM_IMAGEFX_DEINTERLACE_DOUBLE,
'deinterlace2': mmal.MMAL_PARAM_IMAGEFX_DEINTERLACE_ADV,
}
DRC_STRENGTHS = {
'off': mmal.MMAL_PARAMETER_DRC_STRENGTH_OFF,
'low': mmal.MMAL_PARAMETER_DRC_STRENGTH_LOW,
'medium': mmal.MMAL_PARAMETER_DRC_STRENGTH_MEDIUM,
'high': mmal.MMAL_PARAMETER_DRC_STRENGTH_HIGH,
}
RAW_FORMATS = {
'yuv',
'rgb',
'rgba',
'bgr',
'bgra',
}
STEREO_MODES = {
'none': mmal.MMAL_STEREOSCOPIC_MODE_NONE,
'side-by-side': mmal.MMAL_STEREOSCOPIC_MODE_SIDE_BY_SIDE,
'top-bottom': mmal.MMAL_STEREOSCOPIC_MODE_BOTTOM,
}
CLOCK_MODES = {
'reset': mmal.MMAL_PARAM_TIMESTAMP_MODE_RESET_STC,
'raw': mmal.MMAL_PARAM_TIMESTAMP_MODE_RAW_STC,
}
_METER_MODES_R = {v: k for (k, v) in METER_MODES.items()}
_EXPOSURE_MODES_R = {v: k for (k, v) in EXPOSURE_MODES.items()}
_FLASH_MODES_R = {v: k for (k, v) in FLASH_MODES.items()}
_AWB_MODES_R = {v: k for (k, v) in AWB_MODES.items()}
_IMAGE_EFFECTS_R = {v: k for (k, v) in IMAGE_EFFECTS.items()}
_DRC_STRENGTHS_R = {v: k for (k, v) in DRC_STRENGTHS.items()}
_STEREO_MODES_R = {v: k for (k, v) in STEREO_MODES.items()}
_CLOCK_MODES_R = {v: k for (k, v) in CLOCK_MODES.items()}
__slots__ = (
'_used_led',
'_led_pin',
'_camera',
'_camera_config',
'_camera_exception',
'_revision',
'_preview',
'_preview_alpha',
'_preview_layer',
'_preview_fullscreen',
'_preview_window',
'_splitter',
'_splitter_connection',
'_encoders_lock',
'_encoders',
'_overlays',
'_raw_format',
'_image_effect_params',
'_exif_tags',
)
def __init__(
self, camera_num=0, stereo_mode='none', stereo_decimate=False,
resolution=None, framerate=None, sensor_mode=0, led_pin=None,
clock_mode='reset', framerate_range=None):
bcm_host.bcm_host_init()
mimetypes.add_type('application/h264', '.h264', False)
mimetypes.add_type('application/mjpeg', '.mjpg', False)
mimetypes.add_type('application/mjpeg', '.mjpeg', False)
self._used_led = False
if GPIO and led_pin is None:
try:
led_pin = {
(0, 0): 2, # compute module (default for cam 0)
(0, 1): 30, # compute module (default for cam 1)
(1, 0): 5, # Pi 1 model B rev 1
(2, 0): 5, # Pi 1 model B rev 2 or model A
(3, 0): 32, # Pi 1 model B+ or Pi 2 model B
}[(GPIO.RPI_REVISION, camera_num)]
except KeyError:
raise PiCameraError(
'Unable to determine default GPIO LED pin for RPi '
'revision %d and camera num %d' % (
GPIO.RPI_REVISION, camera_num))
self._led_pin = led_pin
self._camera = None
self._camera_config = None
self._camera_exception = None
self._preview = None
self._preview_alpha = 255
self._preview_layer = 2
self._preview_fullscreen = True
self._preview_window = None
self._splitter = None
self._splitter_connection = None
self._encoders_lock = threading.Lock()
self._encoders = {}
self._overlays = []
self._raw_format = 'yuv'
self._image_effect_params = None
with mo.MMALCameraInfo() as camera_info:
info = camera_info.control.params[mmal.MMAL_PARAMETER_CAMERA_INFO]
self._revision = 'ov5647'
if camera_info.info_rev > 1:
self._revision = info.cameras[camera_num].camera_name.decode('ascii')
self._exif_tags = {
'IFD0.Model': 'RP_%s' % self._revision,
'IFD0.Make': 'RaspberryPi',
}
if PiCamera.MAX_RESOLUTION is PiCameraMaxResolution:
PiCamera.MAX_RESOLUTION = mo.PiResolution(
info.cameras[camera_num].max_width,
info.cameras[camera_num].max_height,
)
if PiCamera.MAX_FRAMERATE is PiCameraMaxFramerate:
if self._revision.upper() == 'OV5647':
PiCamera.MAX_FRAMERATE = 90
else:
PiCamera.MAX_FRAMERATE = 120
if resolution is None:
# Get screen resolution
w = ct.c_uint32()
h = ct.c_uint32()
if bcm_host.graphics_get_display_size(0, w, h) == -1:
w = 1280
h = 720
else:
w = int(w.value)
h = int(h.value)
resolution = mo.PiResolution(w, h)
elif resolution is PiCameraMaxResolution:
resolution = PiCamera.MAX_RESOLUTION
else:
resolution = mo.to_resolution(resolution)
if framerate_range is None:
if framerate is None:
framerate = 30
elif framerate is PiCameraMaxFramerate:
framerate = PiCamera.MAX_FRAMERATE
else:
framerate = mo.to_fraction(framerate)
elif framerate is not None:
raise PiCameraValueError(
"Can't specify framerate and framerate_range")
else:
try:
low, high = framerate_range
except TypeError:
raise PiCameraValueError(
"framerate_range must have (low, high) values")
if low is PiCameraMaxFramerate:
low = PiCamera.MAX_FRAMERATE
if high is PiCameraMaxFramerate:
high = PiCamera.MAX_FRAMERATE
framerate = (mo.to_fraction(low), mo.to_fraction(high))
try:
stereo_mode = self.STEREO_MODES[stereo_mode]
except KeyError:
raise PiCameraValueError('Invalid stereo mode: %s' % stereo_mode)
try:
clock_mode = self.CLOCK_MODES[clock_mode]
except KeyError:
raise PiCameraValueError('Invalid clock mode: %s' % clock_mode)
try:
self._init_camera(camera_num, stereo_mode, stereo_decimate)
self._configure_camera(sensor_mode, framerate, resolution, clock_mode)
self._init_preview()
self._init_splitter()
self._camera.enable()
self._init_defaults()
except:
self.close()
raise
def _init_led(self):
global GPIO
if GPIO:
try:
GPIO.setmode(GPIO.BCM)
GPIO.setwarnings(False)
GPIO.setup(self._led_pin, GPIO.OUT, initial=GPIO.LOW)
self._used_led = True
except RuntimeError:
# We're probably not running as root. In this case, forget the
# GPIO reference so we don't try anything further
GPIO = None
def _init_camera(self, num, stereo_mode, stereo_decimate):
try:
self._camera = mo.MMALCamera()
except PiCameraMMALError as e:
if e.status == mmal.MMAL_ENOMEM:
raise PiCameraError(
"Camera is not enabled. Try running 'sudo raspi-config' "
"and ensure that the camera has been enabled.")
else:
raise
self._camera_config = self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_CONFIG]
# Don't attempt to set this if stereo mode isn't requested as it'll
# break compatibility on older firmwares
if stereo_mode != mmal.MMAL_STEREOSCOPIC_MODE_NONE:
for p in self._camera.outputs:
mp = mmal.MMAL_PARAMETER_STEREOSCOPIC_MODE_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_STEREOSCOPIC_MODE,
ct.sizeof(mmal.MMAL_PARAMETER_STEREOSCOPIC_MODE_T),
),
mode=stereo_mode,
decimate=stereo_decimate,
swap_eyes=False,
)
p.params[mmal.MMAL_PARAMETER_STEREOSCOPIC_MODE] = mp
# Must be done *after* stereo-scopic setting
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_NUM] = num
def _init_defaults(self):
self.sharpness = 0
self.contrast = 0
self.brightness = 50
self.saturation = 0
self.iso = 0 # auto
self.video_stabilization = False
self.exposure_compensation = 0
self.exposure_mode = 'auto'
self.meter_mode = 'average'
self.awb_mode = 'auto'
self.image_effect = 'none'
self.color_effects = None
self.rotation = 0
self.hflip = self.vflip = False
self.zoom = (0.0, 0.0, 1.0, 1.0)
def _init_splitter(self):
# Create a splitter component for the video port. This is to permit
# video recordings and captures where use_video_port=True to occur
# simultaneously (#26)
self._splitter = mo.MMALSplitter()
self._splitter.inputs[0].connect(
self._camera.outputs[self.CAMERA_VIDEO_PORT]).enable()
def _init_preview(self):
# Create a null-sink component, enable it and connect it to the
# camera's preview port. If nothing is connected to the preview port,
# the camera doesn't measure exposure and captured images gradually
# fade to black (issue #22)
self._preview = PiNullSink(
self, self._camera.outputs[self.CAMERA_PREVIEW_PORT])
def _start_capture(self, port):
# Only enable capture if the port is the camera's still port, or if
# there's a single active encoder on the video splitter
if (
port == self._camera.outputs[self.CAMERA_CAPTURE_PORT] or
len([e for e in self._encoders.values() if e.active]) == 1):
port.params[mmal.MMAL_PARAMETER_CAPTURE] = True
def _stop_capture(self, port):
# Only disable capture if the port is the camera's still port, or if
# there's a single active encoder on the video splitter
if (
port == self._camera.outputs[self.CAMERA_CAPTURE_PORT] or
len([e for e in self._encoders.values() if e.active]) == 1):
port.params[mmal.MMAL_PARAMETER_CAPTURE] = False
def _check_camera_open(self):
"""
Raise an exception if the camera is already closed, or if the camera
has encountered a fatal error.
"""
exc, self._camera_exception = self._camera_exception, None
if exc:
raise exc
if self.closed:
raise PiCameraClosed("Camera is closed")
def _check_recording_stopped(self):
"""
Raise an exception if the camera is currently recording.
"""
if self.recording:
raise PiCameraRuntimeError("Recording is currently running")
def _get_ports(self, from_video_port, splitter_port):
"""
Determine the camera and output ports for given capture options.
See :ref:`camera_hardware` for more information on picamera's usage of
camera, splitter, and encoder ports. The general idea here is that the
capture (still) port operates on its own, while the video port is
always connected to a splitter component, so requests for a video port
also have to specify which splitter port they want to use.
"""
self._check_camera_open()
if from_video_port and (splitter_port in self._encoders):
raise PiCameraAlreadyRecording(
'The camera is already using port %d ' % splitter_port)
camera_port = (
self._camera.outputs[self.CAMERA_VIDEO_PORT]
if from_video_port else
self._camera.outputs[self.CAMERA_CAPTURE_PORT]
)
output_port = (
self._splitter.outputs[splitter_port]
if from_video_port else
camera_port
)
return (camera_port, output_port)
def _get_output_format(self, output):
"""
Given an output object, attempt to determine the requested format.
We attempt to determine the filename of the *output* object and derive
a MIME type from the extension. If *output* has no filename, an error
is raised.
"""
if isinstance(output, bytes):
filename = output.decode('utf-8')
elif isinstance(output, str):
filename = output
else:
try:
filename = output.name
except AttributeError:
raise PiCameraValueError(
'Format must be specified when output has no filename')
(type, encoding) = mimetypes.guess_type(filename, strict=False)
if not type:
raise PiCameraValueError(
'Unable to determine type from filename %s' % filename)
return type
def _get_image_format(self, output, format=None):
"""
Given an output object and an optional format, attempt to determine the
requested image format.
This method is used by all capture methods to determine the requested
output format. If *format* is specified as a MIME-type the "image/"
prefix is stripped. If *format* is not specified, then
:meth:`_get_output_format` will be called to attempt to determine
format from the *output* object.
"""
if isinstance(format, bytes):
format = format.decode('utf-8')
format = format or self._get_output_format(output)
format = (
format[6:] if format.startswith('image/') else
format)
if format == 'x-ms-bmp':
format = 'bmp'
if format == 'raw':
format = self.raw_format
return format
def _get_video_format(self, output, format=None):
"""
Given an output object and an optional format, attempt to determine the
requested video format.
This method is used by all recording methods to determine the requested
output format. If *format* is specified as a MIME-type the "video/" or
"application/" prefix will be stripped. If *format* is not specified,
then :meth:`_get_output_format` will be called to attempt to determine
format from the *output* object.
"""
if isinstance(format, bytes):
format = format.decode('utf-8')
format = format or self._get_output_format(output)
format = (
format[6:] if format.startswith('video/') else
format[12:] if format.startswith('application/') else
format)
return format
def _get_image_encoder(
self, camera_port, output_port, format, resize, **options):
"""
Construct an image encoder for the requested parameters.
This method is called by :meth:`capture` and :meth:`capture_continuous`
to construct an image encoder. The *camera_port* parameter gives the
MMAL camera port that should be enabled for capture by the encoder. The
*output_port* parameter gives the MMAL port that the encoder should
read output from (this may be the same as the camera port, but may be
different if other component(s) like a splitter have been placed in the
pipeline). The *format* parameter indicates the image format and will
be one of:
* ``'jpeg'``
* ``'png'``
* ``'gif'``
* ``'bmp'``
* ``'yuv'``
* ``'rgb'``
* ``'rgba'``
* ``'bgr'``
* ``'bgra'``
The *resize* parameter indicates the size that the encoder should
resize the output to (presumably by including a resizer in the
pipeline). Finally, *options* includes extra keyword arguments that
should be passed verbatim to the encoder.
"""
encoder_class = (
PiRawOneImageEncoder if format in self.RAW_FORMATS else
PiCookedOneImageEncoder)
return encoder_class(
self, camera_port, output_port, format, resize, **options)
def _get_images_encoder(
self, camera_port, output_port, format, resize, **options):
"""
Construct a multi-image encoder for the requested parameters.
This method is largely equivalent to :meth:`_get_image_encoder` with
the exception that the encoder returned should expect to be passed an
iterable of outputs to its :meth:`~PiEncoder.start` method, rather than
a single output object. This method is called by the
:meth:`capture_sequence` method.
All parameters are the same as in :meth:`_get_image_encoder`. Please
refer to the documentation for that method for further information.
"""
encoder_class = (
PiRawMultiImageEncoder if format in self.RAW_FORMATS else
PiCookedMultiImageEncoder)
return encoder_class(
self, camera_port, output_port, format, resize, **options)
def _get_video_encoder(
self, camera_port, output_port, format, resize, **options):
"""
Construct a video encoder for the requested parameters.
This method is called by :meth:`start_recording` and
:meth:`record_sequence` to construct a video encoder. The
*camera_port* parameter gives the MMAL camera port that should be
enabled for capture by the encoder. The *output_port* parameter gives
the MMAL port that the encoder should read output from (this may be the
same as the camera port, but may be different if other component(s)
like a splitter have been placed in the pipeline). The *format*
parameter indicates the video format and will be one of:
* ``'h264'``
* ``'mjpeg'``
The *resize* parameter indicates the size that the encoder should
resize the output to (presumably by including a resizer in the
pipeline). Finally, *options* includes extra keyword arguments that
should be passed verbatim to the encoder.
"""
encoder_class = (
PiRawVideoEncoder if format in self.RAW_FORMATS else
PiCookedVideoEncoder)
return encoder_class(
self, camera_port, output_port, format, resize, **options)
def close(self):
"""
Finalizes the state of the camera.
After successfully constructing a :class:`PiCamera` object, you should
ensure you call the :meth:`close` method once you are finished with the
camera (e.g. in the ``finally`` section of a ``try..finally`` block).
This method stops all recording and preview activities and releases all
resources associated with the camera; this is necessary to prevent GPU
memory leaks.
"""
for port in list(self._encoders):
self.stop_recording(splitter_port=port)
assert not self.recording
for overlay in list(self._overlays):
self.remove_overlay(overlay)
if self._preview:
self._preview.close()
self._preview = None
if self._splitter:
self._splitter.close()
self._splitter = None
if self._camera:
self._camera.close()
self._camera = None
exc, self._camera_exception = self._camera_exception, None
if exc:
raise exc
def __enter__(self):
return self
def __exit__(self, exc_type, exc_value, exc_tb):
self.close()
def start_preview(self, **options):
"""
Displays the preview overlay.
This method starts a camera preview as an overlay on the Pi's primary
display (HDMI or composite). A :class:`PiRenderer` instance (more
specifically, a :class:`PiPreviewRenderer`) is constructed with the
keyword arguments captured in *options*, and is returned from the
method (this instance is also accessible from the :attr:`preview`
attribute for as long as the renderer remains active). By default, the
renderer will be opaque and fullscreen.
This means the default preview overrides whatever is currently visible
on the display. More specifically, the preview does not rely on a
graphical environment like X-Windows (it can run quite happily from a
TTY console); it is simply an overlay on the Pi's video output. To stop
the preview and reveal the display again, call :meth:`stop_preview`.
The preview can be started and stopped multiple times during the
lifetime of the :class:`PiCamera` object.
All other camera properties can be modified "live" while the preview is
running (e.g. :attr:`brightness`).
.. note::
Because the default preview typically obscures the screen, ensure
you have a means of stopping a preview before starting one. If the
preview obscures your interactive console you won't be able to
Alt+Tab back to it as the preview isn't in a window. If you are in
an interactive Python session, simply pressing Ctrl+D usually
suffices to terminate the environment, including the camera and its
associated preview.
"""
self._check_camera_open()
self._preview.close()
options.setdefault('layer', self._preview_layer)
options.setdefault('alpha', self._preview_alpha)
options.setdefault('fullscreen', self._preview_fullscreen)
options.setdefault('window', self._preview_window)
renderer = PiPreviewRenderer(
self, self._camera.outputs[self.CAMERA_PREVIEW_PORT], **options)
self._preview = renderer
return renderer
def stop_preview(self):
"""
Hides the preview overlay.
If :meth:`start_preview` has previously been called, this method shuts
down the preview display which generally results in the underlying
display becoming visible again. If a preview is not currently running,
no exception is raised - the method will simply do nothing.
"""
self._check_camera_open()
self._preview.close()
self._preview = PiNullSink(
self, self._camera.outputs[self.CAMERA_PREVIEW_PORT])
def add_overlay(self, source, size=None, format=None, **options):
"""
Adds a static overlay to the preview output.
This method creates a new static overlay using the same rendering
mechanism as the preview. Overlays will appear on the Pi's video
output, but will not appear in captures or video recordings. Multiple
overlays can exist; each call to :meth:`add_overlay` returns a new
:class:`PiOverlayRenderer` instance representing the overlay.
The *source* must be an object that supports the :ref:`buffer protocol
<bufferobjects>` in one of the supported unencoded formats: ``'yuv'``,
``'rgb'``, ``'rgba'``, ``'bgr'``, or ``'bgra'``. The format can
specified explicitly with the optional *format* parameter. If not
specified, the method will attempt to guess the format based on the
length of *source* and the *size* (assuming 3 bytes per pixel for RGB,
and 4 bytes for RGBA).
The optional *size* parameter specifies the size of the source image as
a ``(width, height)`` tuple. If this is omitted or ``None`` then the
size is assumed to be the same as the camera's current
:attr:`resolution`.
The length of *source* must take into account that widths are rounded
up to the nearest multiple of 32, and heights to the nearest multiple
of 16. For example, if *size* is ``(1280, 720)``, and *format* is
``'rgb'``, then *source* must be a buffer with length 1280 × 720 × 3
bytes, or 2,764,800 bytes (because 1280 is a multiple of 32, and 720 is
a multiple of 16 no extra rounding is required). However, if *size* is
``(97, 57)``, and *format* is ``'rgb'`` then *source* must be a buffer
with length 128 × 64 × 3 bytes, or 24,576 bytes (pixels beyond column
97 and row 57 in the source will be ignored).
New overlays default to *layer* 0, whilst the preview defaults to layer
2. Higher numbered layers obscure lower numbered layers, hence new
overlays will be invisible (if the preview is running) by default. You
can make the new overlay visible either by making any existing preview
transparent (with the :attr:`~PiRenderer.alpha` property) or by moving
the overlay into a layer higher than the preview (with the
:attr:`~PiRenderer.layer` property).
All keyword arguments captured in *options* are passed onto the
:class:`PiRenderer` constructor. All camera properties except
:attr:`resolution` and :attr:`framerate` can be modified while overlays
exist. The reason for these exceptions is that the overlay has a static
resolution and changing the camera's mode would require resizing of the
source.
.. warning::
If too many overlays are added, the display output will be disabled
and a reboot will generally be required to restore the display.
Overlays are composited "on the fly". Hence, a real-time constraint
exists wherein for each horizontal line of HDMI output, the content
of all source layers must be fetched, resized, converted, and
blended to produce the output pixels.
If enough overlays exist (where "enough" is a number dependent on
overlay size, display resolution, bus frequency, and several other
factors making it unrealistic to calculate in advance), this
process breaks down and video output fails. One solution is to add
``dispmanx_offline=1`` to ``/boot/config.txt`` to force the use of
an off-screen buffer. Be aware that this requires more GPU memory
and may reduce the update rate.
.. _RGB: https://en.wikipedia.org/wiki/RGB
.. _RGBA: https://en.wikipedia.org/wiki/RGBA_color_space
.. versionadded:: 1.8
.. versionchanged:: 1.13
Added *format* parameter
"""
self._check_camera_open()
renderer = PiOverlayRenderer(self, source, size, format, **options)
self._overlays.append(renderer)
return renderer
def remove_overlay(self, overlay):
"""
Removes a static overlay from the preview output.
This method removes an overlay which was previously created by
:meth:`add_overlay`. The *overlay* parameter specifies the
:class:`PiRenderer` instance that was returned by :meth:`add_overlay`.
.. versionadded:: 1.8
"""
if not overlay in self._overlays:
raise PiCameraValueError(
"The specified overlay is not owned by this instance of "
"PiCamera")
overlay.close()
self._overlays.remove(overlay)
def start_recording(
self, output, format=None, resize=None, splitter_port=1, **options):
"""
Start recording video from the camera, storing it in *output*.
If *output* is a string, it will be treated as a filename for a new
file which the video will be written to. If *output* is not a string,
but is an object with a ``write`` method, it is assumed to be a
file-like object and the video data is appended to it (the
implementation only assumes the object has a ``write()`` method - no
other methods are required but ``flush`` will be called at the end of
recording if it is present). If *output* is not a string, and has no
``write`` method it is assumed to be a writeable object implementing
the buffer protocol. In this case, the video frames will be written
sequentially to the underlying buffer (which must be large enough to
accept all frame data).
If *format* is ``None`` (the default), the method will attempt to guess
the required video format from the extension of *output* (if it's a
string), or from the *name* attribute of *output* (if it has one). In
the case that the format cannot be determined, a
:exc:`PiCameraValueError` will be raised.
If *format* is not ``None``, it must be a string specifying the format
that you want the video output in. The format can be a MIME-type or
one of the following strings:
* ``'h264'`` - Write an H.264 video stream
* ``'mjpeg'`` - Write an M-JPEG video stream
* ``'yuv'`` - Write the raw video data to a file in YUV420 format
* ``'rgb'`` - Write the raw video data to a file in 24-bit RGB format
* ``'rgba'`` - Write the raw video data to a file in 32-bit RGBA format
* ``'bgr'`` - Write the raw video data to a file in 24-bit BGR format
* ``'bgra'`` - Write the raw video data to a file in 32-bit BGRA format
If *resize* is not ``None`` (the default), it must be a two-element
tuple specifying the width and height that the video recording should
be resized to. This is particularly useful for recording video using
the full resolution of the camera sensor (which is not possible in
H.264 without down-sizing the output).
The *splitter_port* parameter specifies the port of the built-in
splitter that the video encoder will be attached to. This defaults to
``1`` and most users will have no need to specify anything different.
If you wish to record multiple (presumably resized) streams
simultaneously, specify a value between ``0`` and ``3`` inclusive for
this parameter, ensuring that you do not specify a port that is
currently in use.
Certain formats accept additional options which can be specified
as keyword arguments. The ``'h264'`` format accepts the following
additional options:
* *profile* - The H.264 profile to use for encoding. Defaults to
'high', but can be one of 'baseline', 'main', 'extended', 'high', or
'constrained'.
* *level* - The `H.264 level`_ to use for encoding. Defaults to '4',
but can be any H.264 level up to '4.2'.
* *intra_period* - The key frame rate (the rate at which I-frames are
inserted in the output). Defaults to ``None``, but can be any 32-bit
integer value representing the number of frames between successive
I-frames. The special value 0 causes the encoder to produce a single
initial I-frame, and then only P-frames subsequently. Note that
:meth:`split_recording` will fail in this mode.
* *intra_refresh* - The key frame format (the way in which I-frames
will be inserted into the output stream). Defaults to ``None``, but
can be one of 'cyclic', 'adaptive', 'both', or 'cyclicrows'.
* *inline_headers* - When ``True``, specifies that the encoder should
output SPS/PPS headers within the stream to ensure GOPs (groups of
pictures) are self describing. This is important for streaming
applications where the client may wish to seek within the stream, and
enables the use of :meth:`split_recording`. Defaults to ``True`` if
not specified.
* *sei* - When ``True``, specifies the encoder should include
"Supplemental Enhancement Information" within the output stream.
Defaults to ``False`` if not specified.
* *sps_timing* - When ``True`` the encoder includes the camera's
framerate in the SPS header. Defaults to ``False`` if not specified.
* *motion_output* - Indicates the output destination for motion vector
estimation data. When ``None`` (the default), motion data is not
output. Otherwise, this can be a filename string, a file-like object,
or a writeable buffer object (as with the *output* parameter).
All encoded formats accept the following additional options:
* *bitrate* - The bitrate at which video will be encoded. Defaults to
17000000 (17Mbps) if not specified. The maximum value depends on the
selected `H.264 level`_ and profile. Bitrate 0 indicates the encoder
should not use bitrate control (the encoder is limited by the quality
only).
* *quality* - Specifies the quality that the encoder should attempt
to maintain. For the ``'h264'`` format, use values between 10 and 40
where 10 is extremely high quality, and 40 is extremely low (20-25 is
usually a reasonable range for H.264 encoding). For the ``mjpeg``
format, use JPEG quality values between 1 and 100 (where higher
values are higher quality). Quality 0 is special and seems to be
a "reasonable quality" default.
* *quantization* - Deprecated alias for *quality*.
.. versionchanged:: 1.0
The *resize* parameter was added, and ``'mjpeg'`` was added as a
recording format
.. versionchanged:: 1.3
The *splitter_port* parameter was added
.. versionchanged:: 1.5
The *quantization* parameter was deprecated in favor of *quality*,
and the *motion_output* parameter was added.
.. versionchanged:: 1.11
Support for buffer outputs was added.
.. _H.264 level: https://en.wikipedia.org/wiki/H.264/MPEG-4_AVC#Levels
"""
if 'quantization' in options:
warnings.warn(
PiCameraDeprecated(
'The quantization option is deprecated; please use '
'quality instead (same value)'))
with self._encoders_lock:
camera_port, output_port = self._get_ports(True, splitter_port)
format = self._get_video_format(output, format)
encoder = self._get_video_encoder(
camera_port, output_port, format, resize, **options)
self._encoders[splitter_port] = encoder
try:
encoder.start(output, options.get('motion_output'))
except Exception as e:
encoder.close()
with self._encoders_lock:
del self._encoders[splitter_port]
raise
def split_recording(self, output, splitter_port=1, **options):
"""
Continue the recording in the specified output; close existing output.
When called, the video encoder will wait for the next appropriate
split point (an inline SPS header), then will cease writing to the
current output (and close it, if it was specified as a filename), and
continue writing to the newly specified *output*.
The *output* parameter is treated as in the :meth:`start_recording`
method (it can be a string, a file-like object, or a writeable
buffer object).
The *motion_output* parameter can be used to redirect the output of the
motion vector data in the same fashion as *output*. If *motion_output*
is ``None`` (the default) then motion vector data will not be
redirected and will continue being written to the output specified by
the *motion_output* parameter given to :meth:`start_recording`.
Alternatively, if you only wish to redirect motion vector data, you can
set *output* to ``None`` and given a new value for *motion_output*.
The *splitter_port* parameter specifies which port of the video
splitter the encoder you wish to change outputs is attached to. This
defaults to ``1`` and most users will have no need to specify anything
different. Valid values are between ``0`` and ``3`` inclusive.
Note that unlike :meth:`start_recording`, you cannot specify format or
other options as these cannot be changed in the middle of recording.
Only the new *output* (and *motion_output*) can be specified.
Furthermore, the format of the recording is currently limited to H264,
and *inline_headers* must be ``True`` when :meth:`start_recording` is
called (this is the default).
.. versionchanged:: 1.3
The *splitter_port* parameter was added
.. versionchanged:: 1.5
The *motion_output* parameter was added
.. versionchanged:: 1.11
Support for buffer outputs was added.
"""
try:
with self._encoders_lock:
encoder = self._encoders[splitter_port]
except KeyError:
raise PiCameraNotRecording(
'There is no recording in progress on '
'port %d' % splitter_port)
else:
encoder.split(output, options.get('motion_output'))
def request_key_frame(self, splitter_port=1):
"""
Request the encoder generate a key-frame as soon as possible.
When called, the video encoder running on the specified *splitter_port*
will attempt to produce a key-frame (full-image frame) as soon as
possible. The *splitter_port* defaults to ``1``. Valid values are
between ``0`` and ``3`` inclusive.
.. note::
This method is only meaningful for recordings encoded in the H264
format as MJPEG produces full frames for every frame recorded.
Furthermore, there's no guarantee that the *next* frame will be
a key-frame; this is simply a request to produce one as soon as
possible after the call.
.. versionadded:: 1.11
"""
try:
with self._encoders_lock:
encoder = self._encoders[splitter_port]
except KeyError:
raise PiCameraNotRecording(
'There is no recording in progress on '
'port %d' % splitter_port)
else:
encoder.request_key_frame()
def wait_recording(self, timeout=0, splitter_port=1):
"""
Wait on the video encoder for timeout seconds.
It is recommended that this method is called while recording to check
for exceptions. If an error occurs during recording (for example out of
disk space) the recording will stop, but an exception will only be
raised when the :meth:`wait_recording` or :meth:`stop_recording`
methods are called.
If ``timeout`` is 0 (the default) the function will immediately return
(or raise an exception if an error has occurred).
The *splitter_port* parameter specifies which port of the video
splitter the encoder you wish to wait on is attached to. This
defaults to ``1`` and most users will have no need to specify anything
different. Valid values are between ``0`` and ``3`` inclusive.
.. versionchanged:: 1.3
The *splitter_port* parameter was added
"""
assert timeout is not None
try:
with self._encoders_lock:
encoder = self._encoders[splitter_port]
except KeyError:
raise PiCameraNotRecording(
'There is no recording in progress on '
'port %d' % splitter_port)
else:
encoder.wait(timeout)
def stop_recording(self, splitter_port=1):
"""
Stop recording video from the camera.
After calling this method the video encoder will be shut down and
output will stop being written to the file-like object specified with
:meth:`start_recording`. If an error occurred during recording and
:meth:`wait_recording` has not been called since the error then this
method will raise the exception.
The *splitter_port* parameter specifies which port of the video
splitter the encoder you wish to stop is attached to. This defaults to
``1`` and most users will have no need to specify anything different.
Valid values are between ``0`` and ``3`` inclusive.
.. versionchanged:: 1.3
The *splitter_port* parameter was added
"""
try:
with self._encoders_lock:
encoder = self._encoders[splitter_port]
except KeyError:
raise PiCameraNotRecording(
'There is no recording in progress on '
'port %d' % splitter_port)
else:
try:
self.wait_recording(0, splitter_port)
finally:
encoder.close()
with self._encoders_lock:
del self._encoders[splitter_port]
def record_sequence(
self, outputs, format='h264', resize=None, splitter_port=1, **options):
"""
Record a sequence of video clips from the camera.
This method accepts a sequence or iterator of *outputs* each of which
must either be a string specifying a filename for output, or a
file-like object with a ``write`` method.
The method acts as an iterator itself, yielding each item of the
sequence in turn. In this way, the caller can control how long to
record to each item by only permitting the loop to continue when ready
to switch to the next output.
The *format*, *splitter_port*, *resize*, and *options* parameters are
the same as in :meth:`start_recording`, but *format* defaults to
``'h264'``. The format is **not** derived from the filenames in
*outputs* by this method.
For example, to record 3 consecutive 10-second video clips, writing the
output to a series of H.264 files named clip01.h264, clip02.h264, and
clip03.h264 one could use the following::
import picamera
with picamera.PiCamera() as camera:
for filename in camera.record_sequence([
'clip01.h264',
'clip02.h264',
'clip03.h264']):
print('Recording to %s' % filename)
camera.wait_recording(10)
Alternatively, a more flexible method of writing the previous example
(which is easier to expand to a large number of output files) is by
using a generator expression as the input sequence::
import picamera
with picamera.PiCamera() as camera:
for filename in camera.record_sequence(
'clip%02d.h264' % i for i in range(3)):
print('Recording to %s' % filename)
camera.wait_recording(10)
More advanced techniques are also possible by utilising infinite
sequences, such as those generated by :func:`itertools.cycle`. In the
following example, recording is switched between two in-memory streams.
Whilst one stream is recording, the other is being analysed. The script
only stops recording when a video recording meets some criteria defined
by the ``process`` function::
import io
import itertools
import picamera
with picamera.PiCamera() as camera:
analyse = None
for stream in camera.record_sequence(
itertools.cycle((io.BytesIO(), io.BytesIO()))):
if analyse is not None:
if process(analyse):
break
analyse.seek(0)
analyse.truncate()
camera.wait_recording(5)
analyse = stream
.. versionadded:: 1.3
"""
with self._encoders_lock:
camera_port, output_port = self._get_ports(True, splitter_port)
format = self._get_video_format('', format)
encoder = self._get_video_encoder(
camera_port, output_port, format, resize, **options)
self._encoders[splitter_port] = encoder
try:
start = True
for output in outputs:
if start:
start = False
encoder.start(output, options.get('motion_output'))
else:
encoder.split(output)
yield output
finally:
try:
encoder.wait(0)
finally:
encoder.close()
with self._encoders_lock:
del self._encoders[splitter_port]
def capture(
self, output, format=None, use_video_port=False, resize=None,
splitter_port=0, bayer=False, **options):
"""
Capture an image from the camera, storing it in *output*.
If *output* is a string, it will be treated as a filename for a new
file which the image will be written to. If *output* is not a string,
but is an object with a ``write`` method, it is assumed to be a
file-like object and the image data is appended to it (the
implementation only assumes the object has a ``write`` method - no
other methods are required but ``flush`` will be called at the end of
capture if it is present). If *output* is not a string, and has no
``write`` method it is assumed to be a writeable object implementing
the buffer protocol. In this case, the image data will be written
directly to the underlying buffer (which must be large enough to accept
the image data).
If *format* is ``None`` (the default), the method will attempt to guess
the required image format from the extension of *output* (if it's a
string), or from the *name* attribute of *output* (if it has one). In
the case that the format cannot be determined, a
:exc:`PiCameraValueError` will be raised.
If *format* is not ``None``, it must be a string specifying the format
that you want the image output in. The format can be a MIME-type or
one of the following strings:
* ``'jpeg'`` - Write a JPEG file
* ``'png'`` - Write a PNG file
* ``'gif'`` - Write a GIF file
* ``'bmp'`` - Write a Windows bitmap file
* ``'yuv'`` - Write the raw image data to a file in YUV420 format
* ``'rgb'`` - Write the raw image data to a file in 24-bit RGB format
* ``'rgba'`` - Write the raw image data to a file in 32-bit RGBA format
* ``'bgr'`` - Write the raw image data to a file in 24-bit BGR format
* ``'bgra'`` - Write the raw image data to a file in 32-bit BGRA format
* ``'raw'`` - Deprecated option for raw captures; the format is taken
from the deprecated :attr:`raw_format` attribute
The *use_video_port* parameter controls whether the camera's image or
video port is used to capture images. It defaults to ``False`` which
means that the camera's image port is used. This port is slow but
produces better quality pictures. If you need rapid capture up to the
rate of video frames, set this to ``True``.
When *use_video_port* is ``True``, the *splitter_port* parameter
specifies the port of the video splitter that the image encoder will be
attached to. This defaults to ``0`` and most users will have no need to
specify anything different. This parameter is ignored when
*use_video_port* is ``False``. See :ref:`mmal` for more information
about the video splitter.
If *resize* is not ``None`` (the default), it must be a two-element
tuple specifying the width and height that the image should be resized
to.
.. warning::
If *resize* is specified, or *use_video_port* is ``True``, Exif
metadata will **not** be included in JPEG output. This is due to an
underlying firmware limitation.
Certain file formats accept additional options which can be specified
as keyword arguments. Currently, only the ``'jpeg'`` encoder accepts
additional options, which are:
* *quality* - Defines the quality of the JPEG encoder as an integer
ranging from 1 to 100. Defaults to 85. Please note that JPEG quality
is not a percentage and `definitions of quality`_ vary widely.
* *restart* - Defines the restart interval for the JPEG encoder as a
number of JPEG MCUs. The actual restart interval used will be a
multiple of the number of MCUs per row in the resulting image.
* *thumbnail* - Defines the size and quality of the thumbnail to embed
in the Exif metadata. Specifying ``None`` disables thumbnail
generation. Otherwise, specify a tuple of ``(width, height,
quality)``. Defaults to ``(64, 48, 35)``.
* *bayer* - If ``True``, the raw bayer data from the camera's sensor
is included in the Exif metadata.
.. note::
The so-called "raw" formats listed above (``'yuv'``, ``'rgb'``,
etc.) do not represent the raw bayer data from the camera's sensor.
Rather they provide access to the image data after GPU processing,
but before format encoding (JPEG, PNG, etc). Currently, the only
method of accessing the raw bayer data is via the *bayer* parameter
described above.
.. versionchanged:: 1.0
The *resize* parameter was added, and raw capture formats can now
be specified directly
.. versionchanged:: 1.3
The *splitter_port* parameter was added, and *bayer* was added as
an option for the ``'jpeg'`` format
.. versionchanged:: 1.11
Support for buffer outputs was added.
.. _definitions of quality: http://photo.net/learn/jpeg/#qual
"""
if format == 'raw':
warnings.warn(
PiCameraDeprecated(
'The "raw" format option is deprecated; specify the '
'required format directly instead ("yuv", "rgb", etc.)'))
if use_video_port and bayer:
raise PiCameraValueError(
'bayer is only valid with still port captures')
if 'burst' in options:
raise PiCameraValueError(
'burst is only valid with capture_sequence or capture_continuous')
with self._encoders_lock:
camera_port, output_port = self._get_ports(use_video_port, splitter_port)
format = self._get_image_format(output, format)
encoder = self._get_image_encoder(
camera_port, output_port, format, resize, **options)
if use_video_port:
self._encoders[splitter_port] = encoder
try:
if bayer:
camera_port.params[mmal.MMAL_PARAMETER_ENABLE_RAW_CAPTURE] = True
encoder.start(output)
# Wait for the callback to set the event indicating the end of
# image capture
if not encoder.wait(self.CAPTURE_TIMEOUT):
raise PiCameraRuntimeError(
'Timed out waiting for capture to end')
finally:
encoder.close()
with self._encoders_lock:
if use_video_port:
del self._encoders[splitter_port]
def capture_sequence(
self, outputs, format='jpeg', use_video_port=False, resize=None,
splitter_port=0, burst=False, bayer=False, **options):
"""
Capture a sequence of consecutive images from the camera.
This method accepts a sequence or iterator of *outputs* each of which
must either be a string specifying a filename for output, or a
file-like object with a ``write`` method, or a writeable buffer object.
For each item in the sequence or iterator of outputs, the camera
captures a single image as fast as it can.
The *format*, *use_video_port*, *splitter_port*, *resize*, and
*options* parameters are the same as in :meth:`capture`, but *format*
defaults to ``'jpeg'``. The format is **not** derived from the
filenames in *outputs* by this method.
If *use_video_port* is ``False`` (the default), the *burst* parameter
can be used to make still port captures faster. Specifically, this
prevents the preview from switching resolutions between captures which
significantly speeds up consecutive captures from the still port. The
downside is that this mode is currently has several bugs; the major
issue is that if captures are performed too quickly some frames will
come back severely underexposed. It is recommended that users avoid the
*burst* parameter unless they absolutely require it and are prepared to
work around such issues.
For example, to capture 3 consecutive images::
import time
import picamera
with picamera.PiCamera() as camera:
camera.start_preview()
time.sleep(2)
camera.capture_sequence([
'image1.jpg',
'image2.jpg',
'image3.jpg',
])
camera.stop_preview()
If you wish to capture a large number of images, a list comprehension
or generator expression can be used to construct the list of filenames
to use::
import time
import picamera
with picamera.PiCamera() as camera:
camera.start_preview()
time.sleep(2)
camera.capture_sequence([
'image%02d.jpg' % i
for i in range(100)
])
camera.stop_preview()
More complex effects can be obtained by using a generator function to
provide the filenames or output objects.
.. versionchanged:: 1.0
The *resize* parameter was added, and raw capture formats can now
be specified directly
.. versionchanged:: 1.3
The *splitter_port* parameter was added
.. versionchanged:: 1.11
Support for buffer outputs was added.
"""
if use_video_port:
if burst:
raise PiCameraValueError(
'burst is only valid with still port captures')
if bayer:
raise PiCameraValueError(
'bayer is only valid with still port captures')
with self._encoders_lock:
camera_port, output_port = self._get_ports(use_video_port, splitter_port)
format = self._get_image_format('', format)
if use_video_port:
encoder = self._get_images_encoder(
camera_port, output_port, format, resize, **options)
self._encoders[splitter_port] = encoder
else:
encoder = self._get_image_encoder(
camera_port, output_port, format, resize, **options)
try:
if use_video_port:
encoder.start(outputs)
encoder.wait()
else:
if burst:
camera_port.params[mmal.MMAL_PARAMETER_CAMERA_BURST_CAPTURE] = True
try:
for output in outputs:
if bayer:
camera_port.params[mmal.MMAL_PARAMETER_ENABLE_RAW_CAPTURE] = True
encoder.start(output)
if not encoder.wait(self.CAPTURE_TIMEOUT):
raise PiCameraRuntimeError(
'Timed out waiting for capture to end')
finally:
if burst:
camera_port.params[mmal.MMAL_PARAMETER_CAMERA_BURST_CAPTURE] = False
finally:
encoder.close()
with self._encoders_lock:
if use_video_port:
del self._encoders[splitter_port]
def capture_continuous(
self, output, format=None, use_video_port=False, resize=None,
splitter_port=0, burst=False, bayer=False, **options):
"""
Capture images continuously from the camera as an infinite iterator.
This method returns an infinite iterator of images captured
continuously from the camera. If *output* is a string, each captured
image is stored in a file named after *output* after substitution of
two values with the :meth:`~str.format` method. Those two values are:
* ``{counter}`` - a simple incrementor that starts at 1 and increases
by 1 for each image taken
* ``{timestamp}`` - a :class:`~datetime.datetime` instance
The table below contains several example values of *output* and the
sequence of filenames those values could produce:
.. tabularcolumns:: |p{80mm}|p{40mm}|p{10mm}|
+--------------------------------------------+--------------------------------------------+-------+
| *output* Value | Filenames | Notes |
+============================================+============================================+=======+
| ``'image{counter}.jpg'`` | image1.jpg, image2.jpg, image3.jpg, ... | |
+--------------------------------------------+--------------------------------------------+-------+
| ``'image{counter:02d}.jpg'`` | image01.jpg, image02.jpg, image03.jpg, ... | |
+--------------------------------------------+--------------------------------------------+-------+
| ``'image{timestamp}.jpg'`` | image2013-10-05 12:07:12.346743.jpg, | (1) |
| | image2013-10-05 12:07:32.498539, ... | |
+--------------------------------------------+--------------------------------------------+-------+
| ``'image{timestamp:%H-%M-%S-%f}.jpg'`` | image12-10-02-561527.jpg, | |
| | image12-10-14-905398.jpg | |
+--------------------------------------------+--------------------------------------------+-------+
| ``'{timestamp:%H%M%S}-{counter:03d}.jpg'`` | 121002-001.jpg, 121013-002.jpg, | (2) |
| | 121014-003.jpg, ... | |
+--------------------------------------------+--------------------------------------------+-------+
1. Note that because timestamp's default output includes colons (:),
the resulting filenames are not suitable for use on Windows. For
this reason (and the fact the default contains spaces) it is
strongly recommended you always specify a format when using
``{timestamp}``.
2. You can use both ``{timestamp}`` and ``{counter}`` in a single
format string (multiple times too!) although this tends to be
redundant.
If *output* is not a string, but has a ``write`` method, it is assumed
to be a file-like object and each image is simply written to this
object sequentially. In this case you will likely either want to write
something to the object between the images to distinguish them, or
clear the object between iterations. If *output* is not a string, and
has no ``write`` method, it is assumed to be a writeable object
supporting the buffer protocol; each image is simply written to the
buffer sequentially.
The *format*, *use_video_port*, *splitter_port*, *resize*, and
*options* parameters are the same as in :meth:`capture`.
If *use_video_port* is ``False`` (the default), the *burst* parameter
can be used to make still port captures faster. Specifically, this
prevents the preview from switching resolutions between captures which
significantly speeds up consecutive captures from the still port. The
downside is that this mode is currently has several bugs; the major
issue is that if captures are performed too quickly some frames will
come back severely underexposed. It is recommended that users avoid the
*burst* parameter unless they absolutely require it and are prepared to
work around such issues.
For example, to capture 60 images with a one second delay between them,
writing the output to a series of JPEG files named image01.jpg,
image02.jpg, etc. one could do the following::
import time
import picamera
with picamera.PiCamera() as camera:
camera.start_preview()
try:
for i, filename in enumerate(
camera.capture_continuous('image{counter:02d}.jpg')):
print(filename)
time.sleep(1)
if i == 59:
break
finally:
camera.stop_preview()
Alternatively, to capture JPEG frames as fast as possible into an
in-memory stream, performing some processing on each stream until
some condition is satisfied::
import io
import time
import picamera
with picamera.PiCamera() as camera:
stream = io.BytesIO()
for foo in camera.capture_continuous(stream, format='jpeg'):
# Truncate the stream to the current position (in case
# prior iterations output a longer image)
stream.truncate()
stream.seek(0)
if process(stream):
break
.. versionchanged:: 1.0
The *resize* parameter was added, and raw capture formats can now
be specified directly
.. versionchanged:: 1.3
The *splitter_port* parameter was added
.. versionchanged:: 1.11
Support for buffer outputs was added.
"""
if use_video_port:
if burst:
raise PiCameraValueError(
'burst is only valid with still port captures')
if bayer:
raise PiCameraValueError(
'bayer is only valid with still port captures')
with self._encoders_lock:
camera_port, output_port = self._get_ports(use_video_port, splitter_port)
format = self._get_image_format(output, format)
encoder = self._get_image_encoder(
camera_port, output_port, format, resize, **options)
if use_video_port:
self._encoders[splitter_port] = encoder
try:
if burst:
camera_port.params[mmal.MMAL_PARAMETER_CAMERA_BURST_CAPTURE] = True
try:
if isinstance(output, bytes):
# If we're fed a bytes string, assume it's UTF-8 encoded
# and convert it to Unicode. Technically this is wrong
# (file-systems use all sorts of encodings), but UTF-8 is a
# reasonable default and this keeps compatibility with
# Python 2 simple although it breaks the edge cases of
# non-UTF-8 encoded bytes strings with non-UTF-8 encoded
# file-systems
output = output.decode('utf-8')
if isinstance(output, str):
counter = 1
while True:
filename = output.format(
counter=counter,
timestamp=datetime.datetime.now(),
)
if bayer:
camera_port.params[mmal.MMAL_PARAMETER_ENABLE_RAW_CAPTURE] = True
encoder.start(filename)
if not encoder.wait(self.CAPTURE_TIMEOUT):
raise PiCameraRuntimeError(
'Timed out waiting for capture to end')
yield filename
counter += 1
else:
while True:
if bayer:
camera_port.params[mmal.MMAL_PARAMETER_ENABLE_RAW_CAPTURE] = True
encoder.start(output)
if not encoder.wait(self.CAPTURE_TIMEOUT):
raise PiCameraRuntimeError(
'Timed out waiting for capture to end')
yield output
finally:
if burst:
camera_port.params[mmal.MMAL_PARAMETER_CAMERA_BURST_CAPTURE] = False
finally:
encoder.close()
with self._encoders_lock:
if use_video_port:
del self._encoders[splitter_port]
@property
def closed(self):
"""
Returns ``True`` if the :meth:`close` method has been called.
"""
return not self._camera
@property
def recording(self):
"""
Returns ``True`` if the :meth:`start_recording` method has been called,
and no :meth:`stop_recording` call has been made yet.
"""
return any(
isinstance(e, PiVideoEncoder) and e.active
for e in self._encoders.values()
)
@property
def previewing(self):
"""
Returns ``True`` if the :meth:`start_preview` method has been called,
and no :meth:`stop_preview` call has been made yet.
.. deprecated:: 1.8
Test whether :attr:`preview` is ``None`` instead.
"""
warnings.warn(
PiCameraDeprecated(
'PiCamera.previewing is deprecated; test PiCamera.preview '
'is not None instead'))
return isinstance(self._preview, PiPreviewRenderer)
@property
def revision(self):
"""
Returns a string representing the revision of the Pi's camera module.
At the time of writing, the string returned is 'ov5647' for the V1
module, and 'imx219' for the V2 module.
"""
return self._revision
@property
def exif_tags(self):
"""
Holds a mapping of the Exif tags to apply to captured images.
.. note::
Please note that Exif tagging is only supported with the ``jpeg``
format.
By default several Exif tags are automatically applied to any images
taken with the :meth:`capture` method: ``IFD0.Make`` (which is set to
``RaspberryPi``), ``IFD0.Model`` (which is set to ``RP_OV5647``), and
three timestamp tags: ``IFD0.DateTime``, ``EXIF.DateTimeOriginal``, and
``EXIF.DateTimeDigitized`` which are all set to the current date and
time just before the picture is taken.
If you wish to set additional Exif tags, or override any of the
aforementioned tags, simply add entries to the exif_tags map before
calling :meth:`capture`. For example::
camera.exif_tags['IFD0.Copyright'] = 'Copyright (c) 2013 Foo Industries'
The Exif standard mandates ASCII encoding for all textual values, hence
strings containing non-ASCII characters will cause an encoding error to
be raised when :meth:`capture` is called. If you wish to set binary
values, use a :func:`bytes` value::
camera.exif_tags['EXIF.UserComment'] = b'Something containing\\x00NULL characters'
.. warning::
Binary Exif values are currently ignored; this appears to be a
libmmal or firmware bug.
You may also specify datetime values, integer, or float values, all of
which will be converted to appropriate ASCII strings (datetime values
are formatted as ``YYYY:MM:DD HH:MM:SS`` in accordance with the Exif
standard).
The currently supported Exif tags are:
+-------+-------------------------------------------------------------+
| Group | Tags |
+=======+=============================================================+
| IFD0, | ImageWidth, ImageLength, BitsPerSample, Compression, |
| IFD1 | PhotometricInterpretation, ImageDescription, Make, Model, |
| | StripOffsets, Orientation, SamplesPerPixel, RowsPerString, |
| | StripByteCounts, Xresolution, Yresolution, |
| | PlanarConfiguration, ResolutionUnit, TransferFunction, |
| | Software, DateTime, Artist, WhitePoint, |
| | PrimaryChromaticities, JPEGInterchangeFormat, |
| | JPEGInterchangeFormatLength, YcbCrCoefficients, |
| | YcbCrSubSampling, YcbCrPositioning, ReferenceBlackWhite, |
| | Copyright |
+-------+-------------------------------------------------------------+
| EXIF | ExposureTime, FNumber, ExposureProgram, |
| | SpectralSensitivity, ISOSpeedRatings, OECF, ExifVersion, |
| | DateTimeOriginal, DateTimeDigitized, |
| | ComponentsConfiguration, CompressedBitsPerPixel, |
| | ShutterSpeedValue, ApertureValue, BrightnessValue, |
| | ExposureBiasValue, MaxApertureValue, SubjectDistance, |
| | MeteringMode, LightSource, Flash, FocalLength, SubjectArea, |
| | MakerNote, UserComment, SubSecTime, SubSecTimeOriginal, |
| | SubSecTimeDigitized, FlashpixVersion, ColorSpace, |
| | PixelXDimension, PixelYDimension, RelatedSoundFile, |
| | FlashEnergy, SpacialFrequencyResponse, |
| | FocalPlaneXResolution, FocalPlaneYResolution, |
| | FocalPlaneResolutionUnit, SubjectLocation, ExposureIndex, |
| | SensingMethod, FileSource, SceneType, CFAPattern, |
| | CustomRendered, ExposureMode, WhiteBalance, |
| | DigitalZoomRatio, FocalLengthIn35mmFilm, SceneCaptureType, |
| | GainControl, Contrast, Saturation, Sharpness, |
| | DeviceSettingDescription, SubjectDistanceRange, |
| | ImageUniqueID |
+-------+-------------------------------------------------------------+
| GPS | GPSVersionID, GPSLatitudeRef, GPSLatitude, GPSLongitudeRef, |
| | GPSLongitude, GPSAltitudeRef, GPSAltitude, GPSTimeStamp, |
| | GPSSatellites, GPSStatus, GPSMeasureMode, GPSDOP, |
| | GPSSpeedRef, GPSSpeed, GPSTrackRef, GPSTrack, |
| | GPSImgDirectionRef, GPSImgDirection, GPSMapDatum, |
| | GPSDestLatitudeRef, GPSDestLatitude, GPSDestLongitudeRef, |
| | GPSDestLongitude, GPSDestBearingRef, GPSDestBearing, |
| | GPSDestDistanceRef, GPSDestDistance, GPSProcessingMethod, |
| | GPSAreaInformation, GPSDateStamp, GPSDifferential |
+-------+-------------------------------------------------------------+
| EINT | InteroperabilityIndex, InteroperabilityVersion, |
| | RelatedImageFileFormat, RelatedImageWidth, |
| | RelatedImageLength |
+-------+-------------------------------------------------------------+
"""
return self._exif_tags
def _set_led(self, value):
if not self._used_led:
self._init_led()
if not GPIO:
raise PiCameraRuntimeError(
"GPIO library not found, or not accessible; please install "
"RPi.GPIO and run the script as root")
GPIO.output(self._led_pin, bool(value))
led = property(None, _set_led, doc="""
Sets the state of the camera's LED via GPIO.
If a GPIO library is available (only RPi.GPIO is currently supported),
and if the python process has the necessary privileges (typically this
means running as root via sudo), this property can be used to set the
state of the camera's LED as a boolean value (``True`` is on, ``False``
is off).
.. note::
This is a write-only property. While it can be used to control the
camera's LED, you cannot query the state of the camera's LED using
this property.
.. note::
At present, the camera's LED cannot be controlled on the Pi 3
(the GPIOs used to control the camera LED were re-routed to GPIO
expander on the Pi 3).
.. warning::
There are circumstances in which the camera firmware may override
an existing LED setting. For example, in the case that the firmware
resets the camera (as can happen with a CSI-2 timeout), the LED may
also be reset. If you wish to guarantee that the LED remain off at
all times, you may prefer to use the ``disable_camera_led`` option
in `config.txt`_ (this has the added advantage that sudo privileges
and GPIO access are not required, at least for LED control).
.. _config.txt: https://www.raspberrypi.org/documentation/configuration/config-txt.md
""")
def _get_raw_format(self):
warnings.warn(
PiCameraDeprecated(
'PiCamera.raw_format is deprecated; use required format '
'directly with capture methods instead'))
return self._raw_format
def _set_raw_format(self, value):
warnings.warn(
PiCameraDeprecated(
'PiCamera.raw_format is deprecated; use required format '
'directly with capture methods instead'))
if value not in self.RAW_FORMATS:
raise PiCameraValueError("Invalid raw format: %s" % value)
self._raw_format = value
raw_format = property(_get_raw_format, _set_raw_format, doc="""
Retrieves or sets the raw format of the camera's ports.
.. deprecated:: 1.0
Please use ``'yuv'`` or ``'rgb'`` directly as a format in the
various capture methods instead.
""")
def _get_timestamp(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_SYSTEM_TIME]
timestamp = property(_get_timestamp, doc="""
Retrieves the system time according to the camera firmware.
The camera's timestamp is a 64-bit integer representing the number of
microseconds since the last system boot. When the camera's
:attr:`clock_mode` is ``'raw'`` the values returned by this attribute
are comparable to those from the :attr:`frame`
:attr:`~PiVideoFrame.timestamp` attribute.
""")
def _get_frame(self):
self._check_camera_open()
for e in self._encoders.values():
try:
return e.frame
except AttributeError:
pass
raise PiCameraRuntimeError(
"Cannot query frame information when camera is not recording")
frame = property(_get_frame, doc="""
Retrieves information about the current frame recorded from the camera.
When video recording is active (after a call to
:meth:`start_recording`), this attribute will return a
:class:`PiVideoFrame` tuple containing information about the current
frame that the camera is recording.
If multiple video recordings are currently in progress (after multiple
calls to :meth:`start_recording` with different values for the
``splitter_port`` parameter), which encoder's frame information is
returned is arbitrary. If you require information from a specific
encoder, you will need to extract it from :attr:`_encoders` explicitly.
Querying this property when the camera is not recording will result in
an exception.
.. note::
There is a small window of time when querying this attribute will
return ``None`` after calling :meth:`start_recording`. If this
attribute returns ``None``, this means that the video encoder has
been initialized, but the camera has not yet returned any frames.
""")
def _disable_camera(self):
"""
An internal method for disabling the camera, e.g. for re-configuration.
This disables the splitter and preview connections (if they exist).
"""
self._splitter.connection.disable()
self._preview.renderer.connection.disable()
self._camera.disable()
def _enable_camera(self):
"""
An internal method for enabling the camera after re-configuration.
This ensures the splitter configuration is consistent, then re-enables
the camera along with the splitter and preview connections.
"""
self._camera.enable()
self._preview.renderer.connection.enable()
self._splitter.connection.enable()
def _configure_splitter(self):
"""
Ensures all splitter output ports have a sensible format (I420) and
buffer sizes.
This method is used to ensure the splitter configuration is sane,
typically after :meth:`_configure_camera` is called.
"""
self._splitter.inputs[0].copy_from(self._camera.outputs[self.CAMERA_VIDEO_PORT])
self._splitter.inputs[0].commit()
def _control_callback(self, port, buf):
try:
if buf.command == mmal.MMAL_EVENT_ERROR:
raise PiCameraRuntimeError(
"No data recevied from sensor. Check all connections, "
"including the SUNNY chip on the camera board")
elif buf.command != mmal.MMAL_EVENT_PARAMETER_CHANGED:
raise PiCameraRuntimeError(
"Received unexpected camera control callback event, 0x%08x" % buf[0].cmd)
except Exception as exc:
# Pass the exception to the main thread; next time
# check_camera_open() is called this will get raised
self._camera_exception = exc
def _configure_camera(
self, sensor_mode, framerate, resolution, clock_mode,
old_sensor_mode=0):
"""
An internal method for setting a new camera mode, framerate,
resolution, and/or clock_mode.
This method is used by the setters of the :attr:`resolution`,
:attr:`framerate`, and :attr:`sensor_mode` properties. It assumes the
camera is currently disabled. The *old_mode* and *new_mode* arguments
are required to ensure correct operation on older firmwares
(specifically that we don't try to set the sensor mode when both old
and new modes are 0 or automatic).
"""
old_cc = mmal.MMAL_PARAMETER_CAMERA_CONFIG_T.from_buffer_copy(self._camera_config)
old_ports = [
(port.framesize, port.framerate, port.params[mmal.MMAL_PARAMETER_FPS_RANGE])
for port in self._camera.outputs
]
if old_sensor_mode != 0 or sensor_mode != 0:
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_CUSTOM_SENSOR_CONFIG] = sensor_mode
if not self._camera.control.enabled:
# Initial setup
self._camera.control.enable(self._control_callback)
preview_resolution = resolution
elif (
self._camera.outputs[self.CAMERA_PREVIEW_PORT].framesize ==
self._camera.outputs[self.CAMERA_VIDEO_PORT].framesize
):
preview_resolution = resolution
else:
preview_resolution = self._camera.outputs[self.CAMERA_PREVIEW_PORT].framesize
try:
try:
fps_low, fps_high = framerate
except TypeError:
fps_low = fps_high = framerate
else:
framerate = 0
fps_range = mmal.MMAL_PARAMETER_FPS_RANGE_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_FPS_RANGE,
ct.sizeof(mmal.MMAL_PARAMETER_FPS_RANGE_T)
),
fps_low=mo.to_rational(fps_low),
fps_high=mo.to_rational(fps_high),
)
cc = self._camera_config
cc.max_stills_w = resolution.width
cc.max_stills_h = resolution.height
cc.stills_yuv422 = 0
cc.one_shot_stills = 1
cc.max_preview_video_w = resolution.width
cc.max_preview_video_h = resolution.height
cc.num_preview_video_frames = max(3, fps_high // 10)
cc.stills_capture_circular_buffer_height = 0
cc.fast_preview_resume = 0
cc.use_stc_timestamp = clock_mode
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_CONFIG] = cc
# Clamp preview resolution to camera's resolution
if (
preview_resolution.width > resolution.width or
preview_resolution.height > resolution.height
):
preview_resolution = resolution
for port in self._camera.outputs:
port.params[mmal.MMAL_PARAMETER_FPS_RANGE] = fps_range
if port.index == self.CAMERA_PREVIEW_PORT:
port.framesize = preview_resolution
else:
port.framesize = resolution
port.framerate = framerate
port.commit()
except:
# If anything goes wrong, restore original resolution and
# framerate otherwise the camera can be left in unusual states
# (camera config not matching ports, etc).
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_CONFIG] = old_cc
self._camera_config = old_cc
for port, (res, fps, fps_range) in zip(self._camera.outputs, old_ports):
port.framesize = res
port.framerate = fps
port.params[mmal.MMAL_PARAMETER_FPS_RANGE] = fps_range
port.commit()
raise
def _get_framerate(self):
self._check_camera_open()
port_num = (
self.CAMERA_VIDEO_PORT
if self._encoders else
self.CAMERA_PREVIEW_PORT
)
return mo.PiCameraFraction(self._camera.outputs[port_num].framerate)
def _set_framerate(self, value):
self._check_camera_open()
self._check_recording_stopped()
value = mo.to_fraction(value, den_limit=256)
if not (0 < value <= self.MAX_FRAMERATE):
raise PiCameraValueError("Invalid framerate: %.2ffps" % value)
sensor_mode = self.sensor_mode
clock_mode = self.CLOCK_MODES[self.clock_mode]
resolution = self.resolution
self._disable_camera()
self._configure_camera(
sensor_mode=sensor_mode, framerate=value, resolution=resolution,
clock_mode=clock_mode)
self._configure_splitter()
self._enable_camera()
framerate = property(_get_framerate, _set_framerate, doc="""\
Retrieves or sets the framerate at which video-port based image
captures, video recordings, and previews will run.
When queried, the :attr:`framerate` property returns the rate at which
the camera's video and preview ports will operate as a
:class:`~fractions.Fraction` instance (which can be easily converted to
an :class:`int` or :class:`float`). If :attr:`framerate_range` has been
set, then :attr:`framerate` will be 0 which indicates that a dynamic
range of framerates is being used.
.. note::
For backwards compatibility, a derivative of the
:class:`~fractions.Fraction` class is actually used which permits
the value to be treated as a tuple of ``(numerator, denominator)``.
Setting and retrieving framerate as a ``(numerator, denominator)``
tuple is deprecated and will be removed in 2.0. Please use a
:class:`~fractions.Fraction` instance instead (which is just as
accurate and also permits direct use with math operators).
When set, the property configures the camera so that the next call to
recording and previewing methods will use the new framerate. Setting
this property implicitly sets :attr:`framerate_range` so that the low
and high values are equal to the new framerate. The framerate can be
specified as an :ref:`int <typesnumeric>`, :ref:`float <typesnumeric>`,
:class:`~fractions.Fraction`, or a ``(numerator, denominator)`` tuple.
For example, the following definitions are all equivalent::
from fractions import Fraction
camera.framerate = 30
camera.framerate = 30 / 1
camera.framerate = Fraction(30, 1)
camera.framerate = (30, 1) # deprecated
The camera must not be closed, and no recording must be active when the
property is set.
.. note::
This attribute, in combination with :attr:`resolution`, determines
the mode that the camera operates in. The actual sensor framerate
and resolution used by the camera is influenced, but not directly
set, by this property. See :attr:`sensor_mode` for more
information.
The initial value of this property can be specified with the
*framerate* parameter in the :class:`PiCamera` constructor, and will
default to 30 if not specified.
""")
def _get_sensor_mode(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_CUSTOM_SENSOR_CONFIG]
def _set_sensor_mode(self, value):
self._check_camera_open()
self._check_recording_stopped()
try:
if not (0 <= value <= 7):
raise PiCameraValueError(
"Invalid sensor mode: %d (valid range 0..7)" % value)
except TypeError:
raise PiCameraValueError("Invalid sensor mode: %s" % value)
sensor_mode = self.sensor_mode
clock_mode = self.CLOCK_MODES[self.clock_mode]
resolution = self.resolution
framerate = Fraction(self.framerate)
if framerate == 0:
framerate = self.framerate_range
self._disable_camera()
self._configure_camera(
old_sensor_mode=sensor_mode, sensor_mode=value,
framerate=framerate, resolution=resolution,
clock_mode=clock_mode)
self._configure_splitter()
self._enable_camera()
sensor_mode = property(_get_sensor_mode, _set_sensor_mode, doc="""\
Retrieves or sets the input mode of the camera's sensor.
This is an advanced property which can be used to control the camera's
sensor mode. By default, mode 0 is used which allows the camera to
automatically select an input mode based on the requested
:attr:`resolution` and :attr:`framerate`. Valid values are currently
between 0 and 7. The set of valid sensor modes (along with the
heuristic used to select one automatically) are detailed in the
:ref:`camera_modes` section of the documentation.
.. note::
At the time of writing, setting this property does nothing unless
the camera has been initialized with a sensor mode other than 0.
Furthermore, some mode transitions appear to require setting the
property twice (in a row). This appears to be a firmware
limitation.
The initial value of this property can be specified with the
*sensor_mode* parameter in the :class:`PiCamera` constructor, and will
default to 0 if not specified.
.. versionadded:: 1.9
""")
def _get_clock_mode(self):
self._check_camera_open()
return self._CLOCK_MODES_R[self._camera_config.use_stc_timestamp]
def _set_clock_mode(self, value):
self._check_camera_open()
self._check_recording_stopped()
try:
clock_mode = self.CLOCK_MODES[value]
except KeyError:
raise PiCameraValueError("Invalid clock mode %s" % value)
sensor_mode = self.sensor_mode
framerate = Fraction(self.framerate)
if framerate == 0:
framerate = self.framerate_range
resolution = self.resolution
self._disable_camera()
self._configure_camera(
sensor_mode=sensor_mode, framerate=framerate,
resolution=resolution, clock_mode=clock_mode)
self._configure_splitter()
self._enable_camera()
clock_mode = property(_get_clock_mode, _set_clock_mode, doc="""\
Retrieves or sets the mode of the camera's clock.
This is an advanced property which can be used to control the nature of
the frame timestamps available from the :attr:`frame` property. When
this is "reset" (the default) each frame's timestamp will be relative
to the start of the recording. When this is "raw", each frame's
timestamp will be relative to the last initialization of the camera.
The initial value of this property can be specified with the
*clock_mode* parameter in the :class:`PiCamera` constructor, and will
default to "reset" if not specified.
.. versionadded:: 1.11
""")
def _get_resolution(self):
self._check_camera_open()
return mo.PiResolution(
int(self._camera_config.max_stills_w),
int(self._camera_config.max_stills_h)
)
def _set_resolution(self, value):
self._check_camera_open()
self._check_recording_stopped()
value = mo.to_resolution(value)
if not (
(0 < value.width <= self.MAX_RESOLUTION.width) and
(0 < value.height <= self.MAX_RESOLUTION.height)):
raise PiCameraValueError(
"Invalid resolution requested: %r" % (value,))
sensor_mode = self.sensor_mode
clock_mode = self.CLOCK_MODES[self.clock_mode]
framerate = Fraction(self.framerate)
if framerate == 0:
framerate = self.framerate_range
self._disable_camera()
self._configure_camera(
sensor_mode=sensor_mode, framerate=framerate,
resolution=value, clock_mode=clock_mode)
self._configure_splitter()
self._enable_camera()
resolution = property(_get_resolution, _set_resolution, doc="""
Retrieves or sets the resolution at which image captures, video
recordings, and previews will be captured.
When queried, the :attr:`resolution` property returns the resolution at
which the camera will operate as a tuple of ``(width, height)``
measured in pixels. This is the resolution that the :meth:`capture`
method will produce images at, and the resolution that
:meth:`start_recording` will produce videos at.
When set, the property configures the camera so that the next call to
these methods will use the new resolution. The resolution can be
specified as a ``(width, height)`` tuple, as a string formatted
``'WIDTHxHEIGHT'``, or as a string containing a commonly recognized
`display resolution`_ name (e.g. "VGA", "HD", "1080p", etc). For
example, the following definitions are all equivalent::
camera.resolution = (1280, 720)
camera.resolution = '1280x720'
camera.resolution = '1280 x 720'
camera.resolution = 'HD'
camera.resolution = '720p'
The camera must not be closed, and no recording must be active when the
property is set.
.. note::
This attribute, in combination with :attr:`framerate`, determines
the mode that the camera operates in. The actual sensor framerate
and resolution used by the camera is influenced, but not directly
set, by this property. See :attr:`sensor_mode` for more
information.
The initial value of this property can be specified with the
*resolution* parameter in the :class:`PiCamera` constructor, and will
default to the display's resolution or 1280x720 if the display has
been disabled (with ``tvservice -o``).
.. versionchanged:: 1.11
Resolution permitted to be set as a string. Preview resolution
added as separate property.
.. _display resolution: https://en.wikipedia.org/wiki/Graphics_display_resolution
""")
def _get_framerate_range(self):
self._check_camera_open()
port_num = (
self.CAMERA_VIDEO_PORT
if self._encoders else
self.CAMERA_PREVIEW_PORT
)
mp = self._camera.outputs[port_num].params[mmal.MMAL_PARAMETER_FPS_RANGE]
return mo.PiFramerateRange(
mo.to_fraction(mp.fps_low), mo.to_fraction(mp.fps_high))
def _set_framerate_range(self, value):
self._check_camera_open()
self._check_recording_stopped()
low, high = value
low = mo.to_fraction(low, den_limit=256)
high = mo.to_fraction(high, den_limit=256)
if not (0 < low <= self.MAX_FRAMERATE):
raise PiCameraValueError("Invalid low framerate: %.2ffps" % low)
if not (0 < high <= self.MAX_FRAMERATE):
raise PiCameraValueError("Invalid high framerate: %.2ffps" % high)
if high < low:
raise PiCameraValueError("framerate_range is backwards")
sensor_mode = self.sensor_mode
clock_mode = self.CLOCK_MODES[self.clock_mode]
resolution = self.resolution
self._disable_camera()
self._configure_camera(
sensor_mode=sensor_mode, framerate=(low, high),
resolution=resolution, clock_mode=clock_mode)
self._configure_splitter()
self._enable_camera()
framerate_range = property(_get_framerate_range, _set_framerate_range, doc="""\
Retrieves or sets a range between which the camera's framerate is
allowed to float.
When queried, the :attr:`framerate_range` property returns a
:func:`~collections.namedtuple` derivative with ``low`` and ``high``
components (index 0 and 1 respectively) which specify the limits of the
permitted framerate range.
When set, the property configures the camera so that the next call to
recording and previewing methods will use the new framerate range.
Setting this property will implicitly set the :attr:`framerate`
property to 0 (indicating that a dynamic range of framerates is in use
by the camera).
.. note::
Use of this property prevents use of :attr:`framerate_delta` (there
would be little point in making fractional adjustments to the
framerate when the framerate itself is variable).
The low and high framerates can be specified as :ref:`int
<typesnumeric>`, :ref:`float <typesnumeric>`, or
:class:`~fractions.Fraction` values. For example, the following
definitions are all equivalent::
from fractions import Fraction
camera.framerate_range = (0.16666, 30)
camera.framerate_range = (Fraction(1, 6), 30 / 1)
camera.framerate_range = (Fraction(1, 6), Fraction(30, 1))
The camera must not be closed, and no recording must be active when the
property is set.
.. note::
This attribute, like :attr:`framerate`, determines the mode that
the camera operates in. The actual sensor framerate and resolution
used by the camera is influenced, but not directly set, by this
property. See :attr:`sensor_mode` for more information.
.. versionadded:: 1.13
""")
def _get_framerate_delta(self):
self._check_camera_open()
if self.framerate == 0:
raise PiCameraValueError(
'framerate_delta cannot be used with framerate_range')
port_num = (
self.CAMERA_VIDEO_PORT
if self._encoders else
self.CAMERA_PREVIEW_PORT
)
return self._camera.outputs[port_num].params[mmal.MMAL_PARAMETER_FRAME_RATE] - self.framerate
def _set_framerate_delta(self, value):
self._check_camera_open()
if self.framerate == 0:
raise PiCameraValueError(
'framerate_delta cannot be used with framerate_range')
value = mo.to_fraction(self.framerate + value, den_limit=256)
self._camera.outputs[self.CAMERA_PREVIEW_PORT].params[mmal.MMAL_PARAMETER_FRAME_RATE] = value
self._camera.outputs[self.CAMERA_VIDEO_PORT].params[mmal.MMAL_PARAMETER_FRAME_RATE] = value
framerate_delta = property(_get_framerate_delta, _set_framerate_delta, doc="""\
Retrieves or sets a fractional amount that is added to the camera's
framerate for the purpose of minor framerate adjustments.
When queried, the :attr:`framerate_delta` property returns the amount
that the camera's :attr:`framerate` has been adjusted. This defaults
to 0 (so the camera's framerate is the actual framerate used).
When set, the property adjusts the camera's framerate on the fly. The
property can be set while recordings or previews are in progress. Thus
the framerate used by the camera is actually :attr:`framerate` +
:attr:`framerate_delta`.
.. note::
Framerates deltas can be fractional with adjustments as small as
1/256th of an fps possible (finer adjustments will be rounded).
With an appropriately tuned PID controller, this can be used to
achieve synchronization between the camera framerate and other
devices.
If the new framerate demands a mode switch (such as moving between a
low framerate and a high framerate mode), currently active recordings
may drop a frame. This should only happen when specifying quite large
deltas, or when framerate is at the boundary of a sensor mode (e.g.
49fps).
The framerate delta can be specified as an :ref:`int <typesnumeric>`,
:ref:`float <typesnumeric>`, :class:`~fractions.Fraction` or a
``(numerator, denominator)`` tuple. For example, the following
definitions are all equivalent::
from fractions import Fraction
camera.framerate_delta = 0.5
camera.framerate_delta = 1 / 2 # in python 3
camera.framerate_delta = Fraction(1, 2)
camera.framerate_delta = (1, 2) # deprecated
.. note::
This property is implicitly reset to 0 when :attr:`framerate` or
:attr:`framerate_range` is set. When :attr:`framerate` is 0
(indicating that :attr:`framerate_range` is set), this property
cannot be used. (there would be little point in making fractional
adjustments to the framerate when the framerate itself is
variable).
.. versionadded:: 1.11
""")
def _get_still_stats(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_CAPTURE_STATS_PASS]
def _set_still_stats(self, value):
self._check_camera_open()
self._camera.control.params[mmal.MMAL_PARAMETER_CAPTURE_STATS_PASS] = value
still_stats = property(_get_still_stats, _set_still_stats, doc="""\
Retrieves or sets whether statistics will be calculated from still
frames or the prior preview frame.
When queried, the :attr:`still_stats` property returns a boolean value
indicating when scene statistics will be calculated for still captures
(that is, captures where the *use_video_port* parameter of
:meth:`capture` is ``False``). When this property is ``False`` (the
default), statistics will be calculated from the preceding preview
frame (this also applies when the preview is not visible). When `True`,
statistics will be calculated from the captured image itself.
When set, the propetry controls when scene statistics will be
calculated for still captures. The property can be set while recordings
or previews are in progress. The default value is ``False``.
The advantages to calculating scene statistics from the captured image
are that time between startup and capture is reduced as only the AGC
(automatic gain control) has to converge. The downside is that
processing time for captures increases and that white balance and gain
won't necessarily match the preview.
.. warning::
Enabling the still statistics pass will `override fixed white
balance`_ gains (set via :attr:`awb_gains` and :attr:`awb_mode`).
.. _override fixed white balance: https://www.raspberrypi.org/forums/viewtopic.php?p=875772&sid=92fa4ea70d1fe24590a4cdfb4a10c489#p875772
.. versionadded:: 1.9
""")
def _get_saturation(self):
self._check_camera_open()
return int(self._camera.control.params[mmal.MMAL_PARAMETER_SATURATION] * 100)
def _set_saturation(self, value):
self._check_camera_open()
if not (-100 <= value <= 100):
raise PiCameraValueError(
"Invalid saturation value: %d (valid range -100..100)" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_SATURATION] = Fraction(value, 100)
saturation = property(_get_saturation, _set_saturation, doc="""\
Retrieves or sets the saturation setting of the camera.
When queried, the :attr:`saturation` property returns the color
saturation of the camera as an integer between -100 and 100. When set,
the property adjusts the saturation of the camera. Saturation can be
adjusted while previews or recordings are in progress. The default
value is 0.
""")
def _get_sharpness(self):
self._check_camera_open()
return int(self._camera.control.params[mmal.MMAL_PARAMETER_SHARPNESS] * 100)
def _set_sharpness(self, value):
self._check_camera_open()
if not (-100 <= value <= 100):
raise PiCameraValueError(
"Invalid sharpness value: %d (valid range -100..100)" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_SHARPNESS] = Fraction(value, 100)
sharpness = property(_get_sharpness, _set_sharpness, doc="""\
Retrieves or sets the sharpness setting of the camera.
When queried, the :attr:`sharpness` property returns the sharpness
level of the camera (a measure of the amount of post-processing to
reduce or increase image sharpness) as an integer between -100 and 100.
When set, the property adjusts the sharpness of the camera. Sharpness
can be adjusted while previews or recordings are in progress. The
default value is 0.
""")
def _get_contrast(self):
self._check_camera_open()
return int(self._camera.control.params[mmal.MMAL_PARAMETER_CONTRAST] * 100)
def _set_contrast(self, value):
self._check_camera_open()
if not (-100 <= value <= 100):
raise PiCameraValueError(
"Invalid contrast value: %d (valid range -100..100)" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_CONTRAST] = Fraction(value, 100)
contrast = property(_get_contrast, _set_contrast, doc="""\
Retrieves or sets the contrast setting of the camera.
When queried, the :attr:`contrast` property returns the contrast level
of the camera as an integer between -100 and 100. When set, the
property adjusts the contrast of the camera. Contrast can be adjusted
while previews or recordings are in progress. The default value is 0.
""")
def _get_brightness(self):
self._check_camera_open()
return int(self._camera.control.params[mmal.MMAL_PARAMETER_BRIGHTNESS] * 100)
def _set_brightness(self, value):
self._check_camera_open()
if not (0 <= value <= 100):
raise PiCameraValueError(
"Invalid brightness value: %d (valid range 0..100)" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_BRIGHTNESS] = Fraction(value, 100)
brightness = property(_get_brightness, _set_brightness, doc="""\
Retrieves or sets the brightness setting of the camera.
When queried, the :attr:`brightness` property returns the brightness
level of the camera as an integer between 0 and 100. When set, the
property adjusts the brightness of the camera. Brightness can be
adjusted while previews or recordings are in progress. The default
value is 50.
""")
def _get_shutter_speed(self):
self._check_camera_open()
return int(self._camera.control.params[mmal.MMAL_PARAMETER_SHUTTER_SPEED])
def _set_shutter_speed(self, value):
self._check_camera_open()
self._camera.control.params[mmal.MMAL_PARAMETER_SHUTTER_SPEED] = value
shutter_speed = property(_get_shutter_speed, _set_shutter_speed, doc="""\
Retrieves or sets the shutter speed of the camera in microseconds.
When queried, the :attr:`shutter_speed` property returns the shutter
speed of the camera in microseconds, or 0 which indicates that the
speed will be automatically determined by the auto-exposure algorithm.
Faster shutter times naturally require greater amounts of illumination
and vice versa.
When set, the property adjusts the shutter speed of the camera, which
most obviously affects the illumination of subsequently captured
images. Shutter speed can be adjusted while previews or recordings are
running. The default value is 0 (auto).
.. note::
You can query the :attr:`exposure_speed` attribute to determine the
actual shutter speed being used when this attribute is set to 0.
Please note that this capability requires an up to date firmware
(#692 or later).
.. note::
In later firmwares, this attribute is limited by the value of the
:attr:`framerate` attribute. For example, if framerate is set to
30fps, the shutter speed cannot be slower than 33,333µs (1/fps).
""")
def _get_exposure_speed(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_SETTINGS].exposure
exposure_speed = property(_get_exposure_speed, doc="""\
Retrieves the current shutter speed of the camera.
When queried, this property returns the shutter speed currently being
used by the camera. If you have set :attr:`shutter_speed` to a non-zero
value, then :attr:`exposure_speed` and :attr:`shutter_speed` should be
equal. However, if :attr:`shutter_speed` is set to 0 (auto), then you
can read the actual shutter speed being used from this attribute. The
value is returned as an integer representing a number of microseconds.
This is a read-only property.
.. versionadded:: 1.6
""")
def _get_analog_gain(self):
self._check_camera_open()
return mo.to_fraction(
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_SETTINGS].analog_gain)
analog_gain = property(_get_analog_gain, doc="""\
Retrieves the current analog gain of the camera.
When queried, this property returns the analog gain currently being
used by the camera. The value represents the analog gain of the sensor
prior to digital conversion. The value is returned as a
:class:`~fractions.Fraction` instance.
.. versionadded:: 1.6
""")
def _get_digital_gain(self):
self._check_camera_open()
return mo.to_fraction(
self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_SETTINGS].digital_gain)
digital_gain = property(_get_digital_gain, doc="""\
Retrieves the current digital gain of the camera.
When queried, this property returns the digital gain currently being
used by the camera. The value represents the digital gain the camera
applies after conversion of the sensor's analog output. The value is
returned as a :class:`~fractions.Fraction` instance.
.. versionadded:: 1.6
""")
def _get_video_denoise(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_VIDEO_DENOISE]
def _set_video_denoise(self, value):
self._check_camera_open()
self._camera.control.params[mmal.MMAL_PARAMETER_VIDEO_DENOISE] = value
video_denoise = property(_get_video_denoise, _set_video_denoise, doc="""\
Retrieves or sets whether denoise will be applied to video recordings.
When queried, the :attr:`video_denoise` property returns a boolean
value indicating whether or not the camera software will apply a
denoise algorithm to video recordings.
When set, the property activates or deactivates the denoise algorithm
for video recordings. The property can be set while recordings or
previews are in progress. The default value is ``True``.
.. versionadded:: 1.7
""")
def _get_image_denoise(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_STILLS_DENOISE]
def _set_image_denoise(self, value):
self._check_camera_open()
self._camera.control.params[mmal.MMAL_PARAMETER_STILLS_DENOISE] = value
image_denoise = property(_get_image_denoise, _set_image_denoise, doc="""\
Retrieves or sets whether denoise will be applied to image captures.
When queried, the :attr:`image_denoise` property returns a boolean
value indicating whether or not the camera software will apply a
denoise algorithm to image captures.
When set, the property activates or deactivates the denoise algorithm
for image captures. The property can be set while recordings or
previews are in progress. The default value is ``True``.
.. versionadded:: 1.7
""")
def _get_drc_strength(self):
self._check_camera_open()
return self._DRC_STRENGTHS_R[
self._camera.control.params[mmal.MMAL_PARAMETER_DYNAMIC_RANGE_COMPRESSION].strength
]
def _set_drc_strength(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_DYNAMIC_RANGE_COMPRESSION]
mp.strength = self.DRC_STRENGTHS[value]
except KeyError:
raise PiCameraValueError(
"Invalid dynamic range compression strength: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_DYNAMIC_RANGE_COMPRESSION] = mp
drc_strength = property(_get_drc_strength, _set_drc_strength, doc="""\
Retrieves or sets the dynamic range compression strength of the camera.
When queried, the :attr:`drc_strength` property returns a string
indicating the amount of `dynamic range compression`_ the camera
applies to images.
When set, the attributes adjusts the strength of the dynamic range
compression applied to the camera's output. Valid values are given
in the list below:
{values}
The default value is ``'off'``. All possible values for the attribute
can be obtained from the ``PiCamera.DRC_STRENGTHS`` attribute.
.. warning::
Enabling DRC will `override fixed white balance`_ gains (set via
:attr:`awb_gains` and :attr:`awb_mode`).
.. _dynamic range compression: https://en.wikipedia.org/wiki/Gain_compression
.. _override fixed white balance: https://www.raspberrypi.org/forums/viewtopic.php?p=875772&sid=92fa4ea70d1fe24590a4cdfb4a10c489#p875772
.. versionadded:: 1.6
""".format(values=docstring_values(DRC_STRENGTHS)))
def _get_ISO(self):
warnings.warn(
PiCameraDeprecated(
'PiCamera.ISO is deprecated; use PiCamera.iso instead'))
return self.iso
def _set_ISO(self, value):
warnings.warn(
PiCameraDeprecated(
'PiCamera.ISO is deprecated; use PiCamera.iso instead'))
self.iso = value
ISO = property(_get_ISO, _set_ISO, doc="""
Retrieves or sets the apparent ISO setting of the camera.
.. deprecated:: 1.8
Please use the :attr:`iso` attribute instead.
""")
def _get_iso(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_ISO]
def _set_iso(self, value):
self._check_camera_open()
try:
if not (0 <= value <= 1600):
raise PiCameraValueError(
"Invalid iso value: %d (valid range 0..800)" % value)
except TypeError:
raise PiCameraValueError("Invalid iso value: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_ISO] = value
iso = property(_get_iso, _set_iso, doc="""\
Retrieves or sets the apparent ISO setting of the camera.
When queried, the :attr:`iso` property returns the ISO setting of the
camera, a value which represents the `sensitivity of the camera to
light`_. Lower values (e.g. 100) imply less sensitivity than higher
values (e.g. 400 or 800). Lower sensitivities tend to produce less
"noisy" (smoother) images, but operate poorly in low light conditions.
When set, the property adjusts the sensitivity of the camera (by
adjusting the :attr:`analog_gain` and :attr:`digital_gain`). Valid
values are between 0 (auto) and 1600. The actual value used when iso is
explicitly set will be one of the following values (whichever is
closest): 100, 200, 320, 400, 500, 640, 800.
On the V1 camera module, non-zero ISO values attempt to fix overall
gain at various levels. For example, ISO 100 attempts to provide an
overall gain of 1.0, ISO 200 attempts to provide overall gain of 2.0,
etc. The algorithm prefers analog gain over digital gain to reduce
noise.
On the V2 camera module, ISO 100 attempts to produce overall gain of
~1.84, and ISO 800 attempts to produce overall gain of ~14.72 (the V2
camera module was calibrated against the `ISO film speed`_ standard).
The attribute can be adjusted while previews or recordings are in
progress. The default value is 0 which means automatically determine a
value according to image-taking conditions.
.. note::
Some users on the Pi camera forum have noted that higher ISO values
than 800 (specifically up to 1600) can be achieved in certain
conditions with :attr:`exposure_mode` set to ``'sports'`` and
:attr:`iso` set to 0. It doesn't appear to be possible to manually
request an ISO setting higher than 800, but the picamera library
will permit settings up to 1600 in case the underlying firmware
permits such settings in particular circumstances.
.. note::
Certain :attr:`exposure_mode` values override the ISO setting. For
example, ``'off'`` fixes :attr:`analog_gain` and
:attr:`digital_gain` entirely, preventing this property from
adjusting them when set.
.. _sensitivity of the camera to light: https://en.wikipedia.org/wiki/Film_speed#Digital
.. _ISO film speed: https://en.wikipedia.org/wiki/Film_speed#Current_system:_ISO
""")
def _get_meter_mode(self):
self._check_camera_open()
return self._METER_MODES_R[
self._camera.control.params[mmal.MMAL_PARAMETER_EXP_METERING_MODE].value
]
def _set_meter_mode(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_EXP_METERING_MODE]
mp.value = self.METER_MODES[value]
except KeyError:
raise PiCameraValueError("Invalid metering mode: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_EXP_METERING_MODE] = mp
meter_mode = property(_get_meter_mode, _set_meter_mode, doc="""\
Retrieves or sets the metering mode of the camera.
When queried, the :attr:`meter_mode` property returns the method by
which the camera `determines the exposure`_ as one of the following
strings:
{values}
When set, the property adjusts the camera's metering mode. All modes
set up two regions: a center region, and an outer region. The major
`difference between each mode`_ is the size of the center region. The
``'backlit'`` mode has the largest central region (30% of the width),
while ``'spot'`` has the smallest (10% of the width).
The property can be set while recordings or previews are in progress.
The default value is ``'average'``. All possible values for the
attribute can be obtained from the ``PiCamera.METER_MODES`` attribute.
.. _determines the exposure: https://en.wikipedia.org/wiki/Metering_mode
.. _difference between each mode: https://www.raspberrypi.org/forums/viewtopic.php?p=565644#p565644
""".format(values=docstring_values(METER_MODES)))
def _get_video_stabilization(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_VIDEO_STABILISATION]
def _set_video_stabilization(self, value):
self._check_camera_open()
self._camera.control.params[mmal.MMAL_PARAMETER_VIDEO_STABILISATION] = value
video_stabilization = property(
_get_video_stabilization, _set_video_stabilization, doc="""\
Retrieves or sets the video stabilization mode of the camera.
When queried, the :attr:`video_stabilization` property returns a
boolean value indicating whether or not the camera attempts to
compensate for motion.
When set, the property activates or deactivates video stabilization.
The property can be set while recordings or previews are in progress.
The default value is ``False``.
.. note::
The built-in video stabilization only accounts for `vertical and
horizontal motion`_, not rotation.
.. _vertical and horizontal motion: https://www.raspberrypi.org/forums/viewtopic.php?p=342667&sid=ec7d95e887ab74a90ffaab87888c48cd#p342667
""")
def _get_exposure_compensation(self):
self._check_camera_open()
return self._camera.control.params[mmal.MMAL_PARAMETER_EXPOSURE_COMP]
def _set_exposure_compensation(self, value):
self._check_camera_open()
try:
if not (-25 <= value <= 25):
raise PiCameraValueError(
"Invalid exposure compensation value: "
"%d (valid range -25..25)" % value)
except TypeError:
raise PiCameraValueError(
"Invalid exposure compensation value: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_EXPOSURE_COMP] = value
exposure_compensation = property(
_get_exposure_compensation, _set_exposure_compensation, doc="""\
Retrieves or sets the exposure compensation level of the camera.
When queried, the :attr:`exposure_compensation` property returns an
integer value between -25 and 25 indicating the exposure level of the
camera. Larger values result in brighter images.
When set, the property adjusts the camera's exposure compensation
level. Each increment represents 1/6th of a stop. Hence setting the
attribute to 6 increases exposure by 1 stop. The property can be set
while recordings or previews are in progress. The default value is 0.
""")
def _get_exposure_mode(self):
self._check_camera_open()
return self._EXPOSURE_MODES_R[
self._camera.control.params[mmal.MMAL_PARAMETER_EXPOSURE_MODE].value
]
def _set_exposure_mode(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_EXPOSURE_MODE]
mp.value = self.EXPOSURE_MODES[value]
except KeyError:
raise PiCameraValueError("Invalid exposure mode: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_EXPOSURE_MODE] = mp
exposure_mode = property(_get_exposure_mode, _set_exposure_mode, doc="""\
Retrieves or sets the exposure mode of the camera.
When queried, the :attr:`exposure_mode` property returns a string
representing the exposure setting of the camera. The possible values
can be obtained from the ``PiCamera.EXPOSURE_MODES`` attribute, and
are as follows:
{values}
When set, the property adjusts the camera's exposure mode. The
property can be set while recordings or previews are in progress. The
default value is ``'auto'``.
.. note::
Exposure mode ``'off'`` is special: this disables the camera's
automatic gain control, fixing the values of :attr:`digital_gain`
and :attr:`analog_gain`.
Please note that these properties are not directly settable
(although they can be influenced by setting :attr:`iso` *prior* to
fixing the gains), and default to low values when the camera is
first initialized. Therefore it is important to let them settle on
higher values before disabling automatic gain control otherwise all
frames captured will appear black.
""".format(values=docstring_values(EXPOSURE_MODES)))
def _get_flash_mode(self):
self._check_camera_open()
return self._FLASH_MODES_R[
self._camera.control.params[mmal.MMAL_PARAMETER_FLASH].value
]
def _set_flash_mode(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_FLASH]
mp.value = self.FLASH_MODES[value]
except KeyError:
raise PiCameraValueError("Invalid flash mode: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_FLASH] = mp
flash_mode = property(_get_flash_mode, _set_flash_mode, doc="""\
Retrieves or sets the flash mode of the camera.
When queried, the :attr:`flash_mode` property returns a string
representing the flash setting of the camera. The possible values can
be obtained from the ``PiCamera.FLASH_MODES`` attribute, and are as
follows:
{values}
When set, the property adjusts the camera's flash mode. The property
can be set while recordings or previews are in progress. The default
value is ``'off'``.
.. note::
You must define which GPIO pins the camera is to use for flash and
privacy indicators. This is done within the `Device Tree
configuration`_ which is considered an advanced topic.
Specifically, you need to define pins ``FLASH_0_ENABLE`` and
optionally ``FLASH_0_INDICATOR`` (for the privacy indicator). More
information can be found in this :ref:`recipe
<flash_configuration>`.
.. _Device Tree configuration: https://www.raspberrypi.org/documentation/configuration/pin-configuration.md
.. versionadded:: 1.10
""".format(values=docstring_values(FLASH_MODES)))
def _get_awb_mode(self):
self._check_camera_open()
return self._AWB_MODES_R[
self._camera.control.params[mmal.MMAL_PARAMETER_AWB_MODE].value
]
def _set_awb_mode(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_AWB_MODE]
mp.value = self.AWB_MODES[value]
except KeyError:
raise PiCameraValueError("Invalid auto-white-balance mode: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_AWB_MODE] = mp
awb_mode = property(_get_awb_mode, _set_awb_mode, doc="""\
Retrieves or sets the auto-white-balance mode of the camera.
When queried, the :attr:`awb_mode` property returns a string
representing the auto white balance setting of the camera. The possible
values can be obtained from the ``PiCamera.AWB_MODES`` attribute, and
are as follows:
{values}
When set, the property adjusts the camera's auto-white-balance mode.
The property can be set while recordings or previews are in progress.
The default value is ``'auto'``.
.. note::
AWB mode ``'off'`` is special: this disables the camera's automatic
white balance permitting manual control of the white balance via
the :attr:`awb_gains` property. However, even with AWB disabled,
some attributes (specifically :attr:`still_stats` and
:attr:`drc_strength`) can cause AWB re-calculations.
""".format(values=docstring_values(AWB_MODES)))
def _get_awb_gains(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_CAMERA_SETTINGS]
return (
mo.to_fraction(mp.awb_red_gain),
mo.to_fraction(mp.awb_blue_gain),
)
def _set_awb_gains(self, value):
self._check_camera_open()
try:
red_gain, blue_gain = value
except (ValueError, TypeError):
red_gain = blue_gain = value
if not (0.0 <= red_gain <= 8.0 and 0.0 <= blue_gain <= 8.0):
raise PiCameraValueError(
"Invalid gain(s) in (%f, %f) (valid range: 0.0-8.0)" % (
red_gain, blue_gain))
mp = mmal.MMAL_PARAMETER_AWB_GAINS_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_CUSTOM_AWB_GAINS,
ct.sizeof(mmal.MMAL_PARAMETER_AWB_GAINS_T)
),
mo.to_rational(red_gain),
mo.to_rational(blue_gain),
)
self._camera.control.params[mmal.MMAL_PARAMETER_CUSTOM_AWB_GAINS] = mp
awb_gains = property(_get_awb_gains, _set_awb_gains, doc="""\
Gets or sets the auto-white-balance gains of the camera.
When queried, this attribute returns a tuple of values representing
the `(red, blue)` balance of the camera. The `red` and `blue` values
are returned :class:`~fractions.Fraction` instances. The values will
be between 0.0 and 8.0.
When set, this attribute adjusts the camera's auto-white-balance gains.
The property can be specified as a single value in which case both red
and blue gains will be adjusted equally, or as a `(red, blue)` tuple.
Values can be specified as an :ref:`int <typesnumeric>`, :ref:`float
<typesnumeric>` or :class:`~fractions.Fraction` and each gain must be
between 0.0 and 8.0. Typical values for the gains are between 0.9 and
1.9. The property can be set while recordings or previews are in
progress.
.. note::
This attribute only has an effect when :attr:`awb_mode` is set to
``'off'``. Also note that even with AWB disabled, some attributes
(specifically :attr:`still_stats` and :attr:`drc_strength`) can
cause AWB re-calculations.
.. versionchanged:: 1.6
Prior to version 1.6, this attribute was write-only.
""")
def _get_image_effect(self):
self._check_camera_open()
return self._IMAGE_EFFECTS_R[
self._camera.control.params[mmal.MMAL_PARAMETER_IMAGE_EFFECT].value
]
def _set_image_effect(self, value):
self._check_camera_open()
try:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_IMAGE_EFFECT]
mp.value = self.IMAGE_EFFECTS[value]
self._image_effect_params = None
except KeyError:
raise PiCameraValueError("Invalid image effect: %s" % value)
self._camera.control.params[mmal.MMAL_PARAMETER_IMAGE_EFFECT] = mp
image_effect = property(_get_image_effect, _set_image_effect, doc="""\
Retrieves or sets the current image effect applied by the camera.
When queried, the :attr:`image_effect` property returns a string
representing the effect the camera will apply to captured video. The
possible values can be obtained from the ``PiCamera.IMAGE_EFFECTS``
attribute, and are as follows:
{values}
When set, the property changes the effect applied by the camera. The
property can be set while recordings or previews are in progress, but
only certain effects work while recording video (notably ``'negative'``
and ``'solarize'``). The default value is ``'none'``.
""".format(values=docstring_values(IMAGE_EFFECTS)))
def _get_image_effect_params(self):
self._check_camera_open()
return self._image_effect_params
def _set_image_effect_params(self, value):
self._check_camera_open()
to_int = lambda x: int(x)
to_byte = lambda x: max(0, min(255, int(x)))
to_bool = lambda x: (0, 1)[bool(x)]
to_8dot8 = lambda x: int(x * 256)
valid_transforms = {
'solarize': [
(to_bool, to_byte, to_byte, to_byte, to_byte),
(to_byte, to_byte, to_byte, to_byte),
(to_bool,),
],
'colorpoint': [
(lambda x: max(0, min(3, int(x))),),
],
'colorbalance': [
(to_8dot8, to_8dot8, to_8dot8, to_8dot8, to_int, to_int),
(to_8dot8, to_8dot8, to_8dot8, to_8dot8),
(to_8dot8, to_8dot8, to_8dot8),
],
'colorswap': [
(to_bool,),
],
'posterise': [
(lambda x: max(2, min(31, int(x))),),
],
'blur': [
(lambda x: max(1, min(2, int(x))),),
],
'film': [
(to_byte, to_byte, to_byte),
],
'watercolor': [
(),
(to_byte, to_byte),
]
}
# Ensure params is a tuple
try:
params = tuple(i for i in value)
except TypeError:
params = (value,)
# Find the parameter combination for the current effect
effect = self.image_effect
param_transforms = [
transforms for transforms in valid_transforms.get(effect, [])
if len(transforms) == len(params)
]
if not param_transforms:
raise PiCameraValueError(
'invalid set of parameters for effect "%s"' % effect)
param_transforms = param_transforms[0]
params = tuple(
transform(p)
for (transform, p) in zip(param_transforms, params)
)
mp = mmal.MMAL_PARAMETER_IMAGEFX_PARAMETERS_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_IMAGE_EFFECT_PARAMETERS,
ct.sizeof(mmal.MMAL_PARAMETER_IMAGEFX_PARAMETERS_T)
),
effect=self.IMAGE_EFFECTS[effect],
num_effect_params=len(params),
effect_parameter=params,
)
self._camera.control.params[mmal.MMAL_PARAMETER_IMAGE_EFFECT_PARAMETERS] = mp
self._image_effect_params = value
image_effect_params = property(
_get_image_effect_params, _set_image_effect_params, doc="""\
Retrieves or sets the parameters for the current :attr:`effect
<image_effect>`.
When queried, the :attr:`image_effect_params` property either returns
``None`` (for effects which have no configurable parameters, or if no
parameters have been configured), or a tuple of numeric values up to
six elements long.
When set, the property changes the parameters of the current
:attr:`effect <image_effect>` as a sequence of numbers, or a single
number. Attempting to set parameters on an effect which does not
support parameters, or providing an incompatible set of parameters for
an effect will raise a :exc:`PiCameraValueError` exception.
The effects which have parameters, and what combinations those
parameters can take is as follows:
.. tabularcolumns:: |p{30mm}|p{25mm}|p{75mm}|
+--------------------+----------------+-----------------------------------------+
| Effect | Parameters | Description |
+====================+================+=========================================+
| ``'solarize'`` | *yuv*, | *yuv* controls whether data is |
| | *x0*, *y1*, | processed as RGB (0) or YUV(1). Input |
| | *y2*, *y3* | values from 0 to *x0* - 1 are remapped |
| | | linearly onto the range 0 to *y0*. |
| | | Values from *x0* to 255 are remapped |
| | | linearly onto the range *y1* to *y2*. |
| +----------------+-----------------------------------------+
| | *x0*, *y0*, | Same as above, but *yuv* defaults to |
| | *y1*, *y2* | 0 (process as RGB). |
| +----------------+-----------------------------------------+
| | *yuv* | Same as above, but *x0*, *y0*, *y1*, |
| | | *y2* default to 128, 128, 128, 0 |
| | | respectively. |
+--------------------+----------------+-----------------------------------------+
| ``'colorpoint'`` | *quadrant* | *quadrant* specifies which quadrant |
| | | of the U/V space to retain chroma |
| | | from: 0=green, 1=red/yellow, 2=blue, |
| | | 3=purple. There is no default; this |
| | | effect does nothing until parameters |
| | | are set. |
+--------------------+----------------+-----------------------------------------+
| ``'colorbalance'`` | *lens*, | *lens* specifies the lens shading |
| | *r*, *g*, *b*, | strength (0.0 to 256.0, where 0.0 |
| | *u*, *v* | indicates lens shading has no effect). |
| | | *r*, *g*, *b* are multipliers for their |
| | | respective color channels (0.0 to |
| | | 256.0). *u* and *v* are offsets added |
| | | to the U/V plane (0 to 255). |
| +----------------+-----------------------------------------+
| | *lens*, | Same as above but *u* are defaulted |
| | *r*, *g*, *b* | to 0. |
| +----------------+-----------------------------------------+
| | *lens*, | Same as above but *g* also defaults to |
| | *r*, *b* | to 1.0. |
+--------------------+----------------+-----------------------------------------+
| ``'colorswap'`` | *dir* | If *dir* is 0, swap RGB to BGR. If |
| | | *dir* is 1, swap RGB to BRG. |
+--------------------+----------------+-----------------------------------------+
| ``'posterise'`` | *steps* | Control the quantization steps for the |
| | | image. Valid values are 2 to 32, and |
| | | the default is 4. |
+--------------------+----------------+-----------------------------------------+
| ``'blur'`` | *size* | Specifies the size of the kernel. Valid |
| | | values are 1 or 2. |
+--------------------+----------------+-----------------------------------------+
| ``'film'`` | *strength*, | *strength* specifies the strength of |
| | *u*, *v* | effect. *u* and *v* are offsets added |
| | | to the U/V plane (0 to 255). |
+--------------------+----------------+-----------------------------------------+
| ``'watercolor'`` | *u*, *v* | *u* and *v* specify offsets to add to |
| | | the U/V plane (0 to 255). |
| +----------------+-----------------------------------------+
| | | No parameters indicates no U/V effect. |
+--------------------+----------------+-----------------------------------------+
.. versionadded:: 1.8
""")
def _get_color_effects(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_COLOUR_EFFECT]
if mp.enable != mmal.MMAL_FALSE:
return (mp.u, mp.v)
else:
return None
def _set_color_effects(self, value):
self._check_camera_open()
if value is None:
enable = mmal.MMAL_FALSE
u = v = 128
else:
enable = mmal.MMAL_TRUE
try:
u, v = value
except (TypeError, ValueError) as e:
raise PiCameraValueError(
"Invalid color effect (u, v) tuple: %s" % value)
if not ((0 <= u <= 255) and (0 <= v <= 255)):
raise PiCameraValueError(
"(u, v) values must be between 0 and 255")
mp = mmal.MMAL_PARAMETER_COLOURFX_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_COLOUR_EFFECT,
ct.sizeof(mmal.MMAL_PARAMETER_COLOURFX_T)
),
enable, u, v
)
self._camera.control.params[mmal.MMAL_PARAMETER_COLOUR_EFFECT] = mp
color_effects = property(_get_color_effects, _set_color_effects, doc="""\
Retrieves or sets the current color effect applied by the camera.
When queried, the :attr:`color_effects` property either returns
``None`` which indicates that the camera is using normal color
settings, or a ``(u, v)`` tuple where ``u`` and ``v`` are integer
values between 0 and 255.
When set, the property changes the color effect applied by the camera.
The property can be set while recordings or previews are in progress.
For example, to make the image black and white set the value to ``(128,
128)``. The default value is ``None``.
""")
def _get_rotation(self):
self._check_camera_open()
return self._camera.outputs[0].params[mmal.MMAL_PARAMETER_ROTATION]
def _set_rotation(self, value):
self._check_camera_open()
try:
value = ((int(value) % 360) // 90) * 90
except ValueError:
raise PiCameraValueError("Invalid rotation angle: %s" % value)
for port in self._camera.outputs:
port.params[mmal.MMAL_PARAMETER_ROTATION] = value
rotation = property(_get_rotation, _set_rotation, doc="""\
Retrieves or sets the current rotation of the camera's image.
When queried, the :attr:`rotation` property returns the rotation
applied to the image. Valid values are 0, 90, 180, and 270.
When set, the property changes the rotation applied to the camera's
input. The property can be set while recordings or previews are in
progress. The default value is ``0``.
""")
def _get_vflip(self):
self._check_camera_open()
return self._camera.outputs[0].params[mmal.MMAL_PARAMETER_MIRROR] in (
mmal.MMAL_PARAM_MIRROR_VERTICAL, mmal.MMAL_PARAM_MIRROR_BOTH)
def _set_vflip(self, value):
self._check_camera_open()
value = {
(False, False): mmal.MMAL_PARAM_MIRROR_NONE,
(True, False): mmal.MMAL_PARAM_MIRROR_VERTICAL,
(False, True): mmal.MMAL_PARAM_MIRROR_HORIZONTAL,
(True, True): mmal.MMAL_PARAM_MIRROR_BOTH,
}[(bool(value), self.hflip)]
for port in self._camera.outputs:
port.params[mmal.MMAL_PARAMETER_MIRROR] = value
vflip = property(_get_vflip, _set_vflip, doc="""\
Retrieves or sets whether the camera's output is vertically flipped.
When queried, the :attr:`vflip` property returns a boolean indicating
whether or not the camera's output is vertically flipped. The property
can be set while recordings or previews are in progress. The default
value is ``False``.
""")
def _get_hflip(self):
self._check_camera_open()
return self._camera.outputs[0].params[mmal.MMAL_PARAMETER_MIRROR] in (
mmal.MMAL_PARAM_MIRROR_HORIZONTAL, mmal.MMAL_PARAM_MIRROR_BOTH)
def _set_hflip(self, value):
self._check_camera_open()
value = {
(False, False): mmal.MMAL_PARAM_MIRROR_NONE,
(True, False): mmal.MMAL_PARAM_MIRROR_VERTICAL,
(False, True): mmal.MMAL_PARAM_MIRROR_HORIZONTAL,
(True, True): mmal.MMAL_PARAM_MIRROR_BOTH,
}[(self.vflip, bool(value))]
for port in self._camera.outputs:
port.params[mmal.MMAL_PARAMETER_MIRROR] = value
hflip = property(_get_hflip, _set_hflip, doc="""\
Retrieves or sets whether the camera's output is horizontally flipped.
When queried, the :attr:`hflip` property returns a boolean indicating
whether or not the camera's output is horizontally flipped. The
property can be set while recordings or previews are in progress. The
default value is ``False``.
""")
def _get_zoom(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_INPUT_CROP]
return (
mp.rect.x / 65535.0,
mp.rect.y / 65535.0,
mp.rect.width / 65535.0,
mp.rect.height / 65535.0,
)
def _set_zoom(self, value):
self._check_camera_open()
try:
x, y, w, h = value
except (TypeError, ValueError) as e:
raise PiCameraValueError(
"Invalid zoom rectangle (x, y, w, h) tuple: %s" % value)
mp = mmal.MMAL_PARAMETER_INPUT_CROP_T(
mmal.MMAL_PARAMETER_HEADER_T(
mmal.MMAL_PARAMETER_INPUT_CROP,
ct.sizeof(mmal.MMAL_PARAMETER_INPUT_CROP_T)
),
mmal.MMAL_RECT_T(
max(0, min(65535, int(65535 * x))),
max(0, min(65535, int(65535 * y))),
max(0, min(65535, int(65535 * w))),
max(0, min(65535, int(65535 * h))),
),
)
self._camera.control.params[mmal.MMAL_PARAMETER_INPUT_CROP] = mp
zoom = property(_get_zoom, _set_zoom, doc="""\
Retrieves or sets the zoom applied to the camera's input.
When queried, the :attr:`zoom` property returns a ``(x, y, w, h)``
tuple of floating point values ranging from 0.0 to 1.0, indicating the
proportion of the image to include in the output (this is also known as
the "Region of Interest" or ROI). The default value is ``(0.0, 0.0,
1.0, 1.0)`` which indicates that everything should be included. The
property can be set while recordings or previews are in progress.
The `zoom` is applied to the processed image, after rotation and rescale.
If rotation has been used, zoom is composed of ``(y, x, h, w)`` instead.
The values `w` and `h` can modify the aspect ratio of the image: use equal
values for `w` and `h` if you want to keep the same the aspect ratio.
""")
def _get_crop(self):
warnings.warn(
PiCameraDeprecated(
'PiCamera.crop is deprecated; use PiCamera.zoom instead'))
return self.zoom
def _set_crop(self, value):
warnings.warn(
PiCameraDeprecated(
'PiCamera.crop is deprecated; use PiCamera.zoom instead'))
self.zoom = value
crop = property(_get_crop, _set_crop, doc="""
Retrieves or sets the zoom applied to the camera's input.
.. deprecated:: 1.8
Please use the :attr:`zoom` attribute instead.
""")
def _get_overlays(self):
self._check_camera_open()
return self._overlays
overlays = property(_get_overlays, doc="""\
Retrieves all active :class:`PiRenderer` overlays.
If no overlays are current active, :attr:`overlays` will return an
empty iterable. Otherwise, it will return an iterable of
:class:`PiRenderer` instances which are currently acting as overlays.
Note that the preview renderer is an exception to this: it is *not*
included as an overlay despite being derived from :class:`PiRenderer`.
.. versionadded:: 1.8
""")
def _get_preview(self):
self._check_camera_open()
if isinstance(self._preview, PiPreviewRenderer):
return self._preview
preview = property(_get_preview, doc="""\
Retrieves the :class:`PiRenderer` displaying the camera preview.
If no preview is currently active, :attr:`preview` will return
``None``. Otherwise, it will return the instance of
:class:`PiRenderer` which is currently connected to the camera's
preview port for rendering what the camera sees. You can use the
attributes of the :class:`PiRenderer` class to configure the appearance
of the preview. For example, to make the preview semi-transparent::
import picamera
with picamera.PiCamera() as camera:
camera.start_preview()
camera.preview.alpha = 128
.. versionadded:: 1.8
""")
def _get_preview_alpha(self):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_alpha is deprecated; use '
'PiCamera.preview.alpha instead'))
if self.preview:
return self.preview.alpha
else:
return self._preview_alpha
def _set_preview_alpha(self, value):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_alpha is deprecated; use '
'PiCamera.preview.alpha instead'))
if self.preview:
self.preview.alpha = value
else:
self._preview_alpha = value
preview_alpha = property(_get_preview_alpha, _set_preview_alpha, doc="""\
Retrieves or sets the opacity of the preview window.
.. deprecated:: 1.8
Please use the :attr:`~PiRenderer.alpha` attribute of the
:attr:`preview` object instead.
""")
def _get_preview_layer(self):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_layer is deprecated; '
'use PiCamera.preview.layer instead'))
if self.preview:
return self.preview.layer
else:
return self._preview_layer
def _set_preview_layer(self, value):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_layer is deprecated; '
'use PiCamera.preview.layer instead'))
if self.preview:
self.preview.layer = value
else:
self._preview_layer = value
preview_layer = property(_get_preview_layer, _set_preview_layer, doc="""\
Retrieves or sets the layer of the preview window.
.. deprecated:: 1.8
Please use the :attr:`~PiRenderer.layer` attribute of the
:attr:`preview` object instead.
""")
def _get_preview_fullscreen(self):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_fullscreen is deprecated; '
'use PiCamera.preview.fullscreen instead'))
if self.preview:
return self.preview.fullscreen
else:
return self._preview_fullscreen
def _set_preview_fullscreen(self, value):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_fullscreen is deprecated; '
'use PiCamera.preview.fullscreen instead'))
if self.preview:
self.preview.fullscreen = value
else:
self._preview_fullscreen = value
preview_fullscreen = property(
_get_preview_fullscreen, _set_preview_fullscreen, doc="""\
Retrieves or sets full-screen for the preview window.
.. deprecated:: 1.8
Please use the :attr:`~PiRenderer.fullscreen` attribute of the
:attr:`preview` object instead.
""")
def _get_preview_window(self):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_window is deprecated; '
'use PiCamera.preview.window instead'))
if self.preview:
return self.preview.window
else:
return self._preview_window
def _set_preview_window(self, value):
self._check_camera_open()
warnings.warn(
PiCameraDeprecated(
'PiCamera.preview_window is deprecated; '
'use PiCamera.preview.window instead'))
if self.preview:
self.preview.window = value
else:
self._preview_window = value
preview_window = property(
_get_preview_window, _set_preview_window, doc="""\
Retrieves or sets the size of the preview window.
.. deprecated:: 1.8
Please use the :attr:`~PiRenderer.window` attribute of the
:attr:`preview` object instead.
""")
def _get_annotate_text(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
if mp.enable:
return mp.text.decode('ascii')
else:
return ''
def _set_annotate_text(self, value):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
mp.enable = bool(value or mp.show_frame_num)
if mp.enable:
try:
mp.text = value.encode('ascii')
except ValueError as e:
raise PiCameraValueError(str(e))
self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE] = mp
annotate_text = property(_get_annotate_text, _set_annotate_text, doc="""\
Retrieves or sets a text annotation for all output.
When queried, the :attr:`annotate_text` property returns the current
annotation (if no annotation has been set, this is simply a blank
string).
When set, the property immediately applies the annotation to the
preview (if it is running) and to any future captures or video
recording. Strings longer than 255 characters, or strings containing
non-ASCII characters will raise a :exc:`PiCameraValueError`. The
default value is ``''``.
.. versionchanged:: 1.8
Text annotations can now be 255 characters long. The prior limit
was 32 characters.
""")
def _get_annotate_frame_num(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
return mp.show_frame_num.value != mmal.MMAL_FALSE
def _set_annotate_frame_num(self, value):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
mp.enable = bool(value or mp.text)
mp.show_frame_num = bool(value)
self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE] = mp
annotate_frame_num = property(
_get_annotate_frame_num, _set_annotate_frame_num, doc="""\
Controls whether the current frame number is drawn as an annotation.
The :attr:`annotate_frame_num` attribute is a bool indicating whether
or not the current frame number is rendered as an annotation, similar
to :attr:`annotate_text`. The default is ``False``.
.. versionadded:: 1.8
""")
def _get_annotate_text_size(self):
self._check_camera_open()
if self._camera.annotate_rev == 3:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
return mp.text_size or self.DEFAULT_ANNOTATE_SIZE
else:
return self.DEFAULT_ANNOTATE_SIZE
def _set_annotate_text_size(self, value):
self._check_camera_open()
if not (6 <= value <= 160):
raise PiCameraValueError(
"Invalid annotation text size: %d (valid range 6-160)" % value)
if self._camera.annotate_rev == 3:
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
mp.text_size = value
self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE] = mp
elif value != self.DEFAULT_ANNOTATE_SIZE:
warnings.warn(
PiCameraFallback(
"Firmware does not support setting annotation text "
"size; using default (%d) instead" % self.DEFAULT_ANNOTATE_SIZE))
annotate_text_size = property(
_get_annotate_text_size, _set_annotate_text_size, doc="""\
Controls the size of the annotation text.
The :attr:`annotate_text_size` attribute is an int which determines how
large the annotation text will appear on the display. Valid values are
in the range 6 to 160, inclusive. The default is {size}.
.. versionadded:: 1.10
""".format(size=DEFAULT_ANNOTATE_SIZE))
def _get_annotate_foreground(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
if self._camera.annotate_rev == 3 and mp.custom_text_color:
return Color.from_yuv_bytes(
mp.custom_text_Y,
mp.custom_text_U,
mp.custom_text_V)
else:
return Color('white')
def _set_annotate_foreground(self, value):
self._check_camera_open()
if not isinstance(value, Color):
raise PiCameraValueError(
'annotate_foreground must be a Color')
elif self._camera.annotate_rev < 3:
if value.rgb_bytes != (255, 255, 255):
warnings.warn(
PiCameraFallback(
"Firmware does not support setting a custom foreground "
"annotation color; using white instead"))
return
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
mp.custom_text_color = True
(
mp.custom_text_Y,
mp.custom_text_U,
mp.custom_text_V,
) = value.yuv_bytes
self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE] = mp
annotate_foreground = property(
_get_annotate_foreground, _set_annotate_foreground, doc="""\
Controls the color of the annotation text.
The :attr:`annotate_foreground` attribute specifies, partially, the
color of the annotation text. The value is specified as a
:class:`Color`. The default is white.
.. note::
The underlying firmware does not directly support setting all
components of the text color, only the Y' component of a `Y'UV`_
tuple. This is roughly (but not precisely) analogous to the
"brightness" of a color, so you may choose to think of this as
setting how bright the annotation text will be relative to its
background. In order to specify just the Y' component when setting
this attribute, you may choose to construct the
:class:`Color` instance as follows::
camera.annotate_foreground = picamera.Color(y=0.2, u=0, v=0)
.. _Y'UV: https://en.wikipedia.org/wiki/YUV
.. versionadded:: 1.10
""")
def _get_annotate_background(self):
self._check_camera_open()
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
if self._camera.annotate_rev == 3:
if mp.enable_text_background:
if mp.custom_background_color:
return Color.from_yuv_bytes(
mp.custom_background_Y,
mp.custom_background_U,
mp.custom_background_V)
else:
return Color('black')
else:
return None
else:
if mp.black_text_background:
return Color('black')
else:
return None
def _set_annotate_background(self, value):
self._check_camera_open()
if value is True:
warnings.warn(
PiCameraDeprecated(
'Setting PiCamera.annotate_background to True is '
'deprecated; use PiCamera.color.Color("black") instead'))
value = Color('black')
elif value is False:
warnings.warn(
PiCameraDeprecated(
'Setting PiCamera.annotate_background to False is '
'deprecated; use None instead'))
value = None
elif value is None:
pass
elif not isinstance(value, Color):
raise PiCameraValueError(
'annotate_background must be a Color or None')
elif self._camera.annotate_rev < 3 and value.rgb_bytes != (0, 0, 0):
warnings.warn(
PiCameraFallback(
"Firmware does not support setting a custom background "
"annotation color; using black instead"))
mp = self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE]
if self._camera.annotate_rev == 3:
if value is None:
mp.enable_text_background = False
else:
mp.enable_text_background = True
mp.custom_background_color = True
(
mp.custom_background_Y,
mp.custom_background_U,
mp.custom_background_V,
) = value.yuv_bytes
else:
if value is None:
mp.black_text_background = False
else:
mp.black_text_background = True
self._camera.control.params[mmal.MMAL_PARAMETER_ANNOTATE] = mp
annotate_background = property(
_get_annotate_background, _set_annotate_background, doc="""\
Controls what background is drawn behind the annotation.
The :attr:`annotate_background` attribute specifies if a background
will be drawn behind the :attr:`annotation text <annotate_text>` and,
if so, what color it will be. The value is specified as a
:class:`Color` or ``None`` if no background should be drawn. The
default is ``None``.
.. note::
For backward compatibility purposes, the value ``False`` will be
treated as ``None``, and the value ``True`` will be treated as the
color black. The "truthiness" of the values returned by the
attribute are backward compatible although the values themselves
are not.
.. versionadded:: 1.8
.. versionchanged:: 1.10
In prior versions this was a bool value with ``True`` representing
a black background.
""")
| 45.465936 | 146 | 0.609215 | [
"BSD-3-Clause"
] | RobertLucian/picamera | picamera/camera.py | 174,185 | Python |
# Copyright 2020 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ============================================================================
"""logical_and_impl"""
from mindspore.ops.composite import base
from mindspore.ops import functional as F
# logical_and is a metagraph object which will generate function according to input type
# using ".register" decorator
logical_and = base.MultitypeFuncGraph("logical_and")
@logical_and.register("Number", "Number")
def _logical_and_scala(x, y):
"""
Return logical and operation result of x and y.
Args:
x(Number): Number.
y(Number): Number.
Returns:
bool, Return logical and operation result of x and y.
"""
return F.bool_and(x.__bool__(), y.__bool__())
@logical_and.register("Tensor", "Tensor")
def _logical_and_tensor(x, y):
"""
Return logical and operation result of x and y.
Args:
x(Tensor): Tensor.
y(Tensor): Tensor.
Returns:
Tensor, Return logical and operation result of x and y.
"""
return F.logical_and(x, y)
| 30 | 88 | 0.67673 | [
"Apache-2.0"
] | Gavin-Hoang/mindspore | mindspore/ops/composite/multitype_ops/logical_and_impl.py | 1,590 | Python |
from jsonrpc import ServiceProxy
import sys
import string
# ===== BEGIN USER SETTINGS =====
# if you do not set these you will be prompted for a password for every command
rpcuser = ""
rpcpass = ""
# ====== END USER SETTINGS ======
if rpcpass == "":
access = ServiceProxy("http://127.0.0.1:9332")
else:
access = ServiceProxy("http://"+rpcuser+":"+rpcpass+"@127.0.0.1:9332")
cmd = sys.argv[1].lower()
if cmd == "backupwallet":
try:
path = raw_input("Enter destination path/filename: ")
print access.backupwallet(path)
except:
print "\n---An error occurred---\n"
elif cmd == "getaccount":
try:
addr = raw_input("Enter a Sarnath address: ")
print access.getaccount(addr)
except:
print "\n---An error occurred---\n"
elif cmd == "getaccountaddress":
try:
acct = raw_input("Enter an account name: ")
print access.getaccountaddress(acct)
except:
print "\n---An error occurred---\n"
elif cmd == "getaddressesbyaccount":
try:
acct = raw_input("Enter an account name: ")
print access.getaddressesbyaccount(acct)
except:
print "\n---An error occurred---\n"
elif cmd == "getbalance":
try:
acct = raw_input("Enter an account (optional): ")
mc = raw_input("Minimum confirmations (optional): ")
try:
print access.getbalance(acct, mc)
except:
print access.getbalance()
except:
print "\n---An error occurred---\n"
elif cmd == "getblockbycount":
try:
height = raw_input("Height: ")
print access.getblockbycount(height)
except:
print "\n---An error occurred---\n"
elif cmd == "getblockcount":
try:
print access.getblockcount()
except:
print "\n---An error occurred---\n"
elif cmd == "getblocknumber":
try:
print access.getblocknumber()
except:
print "\n---An error occurred---\n"
elif cmd == "getconnectioncount":
try:
print access.getconnectioncount()
except:
print "\n---An error occurred---\n"
elif cmd == "getdifficulty":
try:
print access.getdifficulty()
except:
print "\n---An error occurred---\n"
elif cmd == "getgenerate":
try:
print access.getgenerate()
except:
print "\n---An error occurred---\n"
elif cmd == "gethashespersec":
try:
print access.gethashespersec()
except:
print "\n---An error occurred---\n"
elif cmd == "getinfo":
try:
print access.getinfo()
except:
print "\n---An error occurred---\n"
elif cmd == "getnewaddress":
try:
acct = raw_input("Enter an account name: ")
try:
print access.getnewaddress(acct)
except:
print access.getnewaddress()
except:
print "\n---An error occurred---\n"
elif cmd == "getreceivedbyaccount":
try:
acct = raw_input("Enter an account (optional): ")
mc = raw_input("Minimum confirmations (optional): ")
try:
print access.getreceivedbyaccount(acct, mc)
except:
print access.getreceivedbyaccount()
except:
print "\n---An error occurred---\n"
elif cmd == "getreceivedbyaddress":
try:
addr = raw_input("Enter a Sarnath address (optional): ")
mc = raw_input("Minimum confirmations (optional): ")
try:
print access.getreceivedbyaddress(addr, mc)
except:
print access.getreceivedbyaddress()
except:
print "\n---An error occurred---\n"
elif cmd == "gettransaction":
try:
txid = raw_input("Enter a transaction ID: ")
print access.gettransaction(txid)
except:
print "\n---An error occurred---\n"
elif cmd == "getwork":
try:
data = raw_input("Data (optional): ")
try:
print access.gettransaction(data)
except:
print access.gettransaction()
except:
print "\n---An error occurred---\n"
elif cmd == "help":
try:
cmd = raw_input("Command (optional): ")
try:
print access.help(cmd)
except:
print access.help()
except:
print "\n---An error occurred---\n"
elif cmd == "listaccounts":
try:
mc = raw_input("Minimum confirmations (optional): ")
try:
print access.listaccounts(mc)
except:
print access.listaccounts()
except:
print "\n---An error occurred---\n"
elif cmd == "listreceivedbyaccount":
try:
mc = raw_input("Minimum confirmations (optional): ")
incemp = raw_input("Include empty? (true/false, optional): ")
try:
print access.listreceivedbyaccount(mc, incemp)
except:
print access.listreceivedbyaccount()
except:
print "\n---An error occurred---\n"
elif cmd == "listreceivedbyaddress":
try:
mc = raw_input("Minimum confirmations (optional): ")
incemp = raw_input("Include empty? (true/false, optional): ")
try:
print access.listreceivedbyaddress(mc, incemp)
except:
print access.listreceivedbyaddress()
except:
print "\n---An error occurred---\n"
elif cmd == "listtransactions":
try:
acct = raw_input("Account (optional): ")
count = raw_input("Number of transactions (optional): ")
frm = raw_input("Skip (optional):")
try:
print access.listtransactions(acct, count, frm)
except:
print access.listtransactions()
except:
print "\n---An error occurred---\n"
elif cmd == "move":
try:
frm = raw_input("From: ")
to = raw_input("To: ")
amt = raw_input("Amount:")
mc = raw_input("Minimum confirmations (optional): ")
comment = raw_input("Comment (optional): ")
try:
print access.move(frm, to, amt, mc, comment)
except:
print access.move(frm, to, amt)
except:
print "\n---An error occurred---\n"
elif cmd == "sendfrom":
try:
frm = raw_input("From: ")
to = raw_input("To: ")
amt = raw_input("Amount:")
mc = raw_input("Minimum confirmations (optional): ")
comment = raw_input("Comment (optional): ")
commentto = raw_input("Comment-to (optional): ")
try:
print access.sendfrom(frm, to, amt, mc, comment, commentto)
except:
print access.sendfrom(frm, to, amt)
except:
print "\n---An error occurred---\n"
elif cmd == "sendmany":
try:
frm = raw_input("From: ")
to = raw_input("To (in format address1:amount1,address2:amount2,...): ")
mc = raw_input("Minimum confirmations (optional): ")
comment = raw_input("Comment (optional): ")
try:
print access.sendmany(frm,to,mc,comment)
except:
print access.sendmany(frm,to)
except:
print "\n---An error occurred---\n"
elif cmd == "sendtoaddress":
try:
to = raw_input("To (in format address1:amount1,address2:amount2,...): ")
amt = raw_input("Amount:")
comment = raw_input("Comment (optional): ")
commentto = raw_input("Comment-to (optional): ")
try:
print access.sendtoaddress(to,amt,comment,commentto)
except:
print access.sendtoaddress(to,amt)
except:
print "\n---An error occurred---\n"
elif cmd == "setaccount":
try:
addr = raw_input("Address: ")
acct = raw_input("Account:")
print access.setaccount(addr,acct)
except:
print "\n---An error occurred---\n"
elif cmd == "setgenerate":
try:
gen= raw_input("Generate? (true/false): ")
cpus = raw_input("Max processors/cores (-1 for unlimited, optional):")
try:
print access.setgenerate(gen, cpus)
except:
print access.setgenerate(gen)
except:
print "\n---An error occurred---\n"
elif cmd == "settxfee":
try:
amt = raw_input("Amount:")
print access.settxfee(amt)
except:
print "\n---An error occurred---\n"
elif cmd == "stop":
try:
print access.stop()
except:
print "\n---An error occurred---\n"
elif cmd == "validateaddress":
try:
addr = raw_input("Address: ")
print access.validateaddress(addr)
except:
print "\n---An error occurred---\n"
elif cmd == "walletpassphrase":
try:
pwd = raw_input("Enter wallet passphrase: ")
access.walletpassphrase(pwd, 60)
print "\n---Wallet unlocked---\n"
except:
print "\n---An error occurred---\n"
elif cmd == "walletpassphrasechange":
try:
pwd = raw_input("Enter old wallet passphrase: ")
pwd2 = raw_input("Enter new wallet passphrase: ")
access.walletpassphrasechange(pwd, pwd2)
print
print "\n---Passphrase changed---\n"
except:
print
print "\n---An error occurred---\n"
print
else:
print "Command not found or not supported"
| 24.110769 | 79 | 0.668198 | [
"MIT"
] | iannkwon/Sarnath | contrib/bitrpc/bitrpc.py | 7,836 | Python |
import matplotlib.pyplot as plt
import matplotlib
import numpy as np
from PIL import Image
from scipy.misc import imsave, imread
def plots(epochs, train_acc, test_acc, train_loss, test_loss, train_error, test_error,filename):
plt.style.use('bmh')
fig=plt.figure(figsize=(8,6))
plt.plot(epochs,train_acc, 'r', epochs,test_acc, 'g')
plt.title('model accuracy')
plt.ylabel('accuracy')
plt.xlabel('epoch')
plt.legend(['train_acc', 'test_acc'], loc='upper left')
fig.savefig(filename + '_accuracy.png')
fig=plt.figure(figsize=(8,6))
plt.plot(epochs,train_loss, 'r', epochs,test_loss, 'g')
plt.title('model loss')
plt.ylabel('loss')
plt.xlabel('epoch')
plt.legend(['train_loss', 'test_loss'], loc='upper left')
fig.savefig(filename + '_loss.png')
fig=plt.figure(figsize=(8,6))
plt.plot(epochs,train_error, 'r', epochs,test_error, 'g')
plt.title('model error rate')
plt.ylabel('error rate')
plt.xlabel('epoch')
plt.legend(['train_error', 'test_error'], loc='upper left')
fig.savefig(filename + '_error.png')
plt.close('all')
def write_csv(filename, train_acc,test_acc,train_loss,test_loss,train_error,test_error,epoch):
if epoch==0:
with open(filename, 'w') as f:
f.write('train_acc,test_acc,train_loss, test_loss, train_error, test_error\n')
f.write('{0},{1},{2},{3},{4},{5}\n'.format(train_acc[-1],\
test_acc[-1],\
train_loss[-1],\
test_loss[-1],\
train_error[-1],\
test_error[-1]))
else:
with open(filename, 'a') as f:
f.write('{0},{1},{2},{3},{4},{5}\n'.format(train_acc[-1],\
test_acc[-1],\
train_loss[-1],\
test_loss[-1],\
train_error[-1],\
test_error[-1]))
| 35.112903 | 96 | 0.509417 | [
"MIT"
] | gahshiv/DenseNet-pytorch | data_utils.py | 2,177 | Python |
###############################################################################
#
# Tests for XlsxWriter.
#
# SPDX-License-Identifier: BSD-2-Clause
# Copyright (c), 2013-2022, John McNamara, [email protected]
#
from ..excel_comparison_test import ExcelComparisonTest
from ...workbook import Workbook
class TestCompareXLSXFiles(ExcelComparisonTest):
"""
Test file created by XlsxWriter against a file created by Excel.
"""
def setUp(self):
self.set_filename('chart_data_labels17.xlsx')
self.ignore_elements = {'xl/charts/chart1.xml': ['<c:formatCode']}
def test_create_file(self):
"""Test the creation of a simple XlsxWriter file."""
workbook = Workbook(self.got_filename)
worksheet = workbook.add_worksheet()
chart = workbook.add_chart({'type': 'stock'})
date_format = workbook.add_format({'num_format': 14})
chart.axis_ids = [45740032, 45747200]
data = [
[39083, 39084, 39085, 39086, 39087],
[27.2, 25.03, 19.05, 20.34, 18.5],
[23.49, 19.55, 15.12, 17.84, 16.34],
[25.45, 23.05, 17.32, 20.45, 17.34],
]
for row in range(5):
worksheet.write(row, 0, data[0][row], date_format)
worksheet.write(row, 1, data[1][row])
worksheet.write(row, 2, data[2][row])
worksheet.write(row, 3, data[3][row])
worksheet.set_column('A:D', 11)
chart.add_series({
'categories': '=Sheet1!$A$1:$A$5',
'values': '=Sheet1!$B$1:$B$5',
})
chart.add_series({
'categories': '=Sheet1!$A$1:$A$5',
'values': '=Sheet1!$C$1:$C$5',
})
chart.add_series({
'categories': '=Sheet1!$A$1:$A$5',
'values': '=Sheet1!$D$1:$D$5',
'data_labels': {'value': 1, 'position': 'right'},
})
worksheet.insert_chart('E9', chart)
workbook.close()
self.assertExcelEqual()
| 27.805556 | 79 | 0.535964 | [
"BSD-2-Clause"
] | hugovk/XlsxWriter | xlsxwriter/test/comparison/test_chart_data_labels17.py | 2,002 | Python |
from datetime import timedelta
from typing import NamedTuple, Optional
class ErdAdvantiumKitchenTimerMinMax(NamedTuple):
"""Defines min/max kitchen timer settings"""
min_time: timedelta
max_time: timedelta
raw_value: Optional[str]
| 24.9 | 49 | 0.7751 | [
"MIT"
] | ChevySSinSD/gehome | gehomesdk/erd/values/advantium/erd_advantium_kitchen_timer_min_max.py | 249 | Python |
#!/usr/bin/env python3
# Copyright (c) 2018-2020 The Ludirium Core developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Useful util functions for testing the wallet"""
from collections import namedtuple
from test_framework.address import (
byte_to_base58,
key_to_p2pkh,
key_to_p2sh_p2wpkh,
key_to_p2wpkh,
script_to_p2sh,
script_to_p2sh_p2wsh,
script_to_p2wsh,
)
from test_framework.key import ECKey
from test_framework.script import (
CScript,
OP_2,
OP_3,
OP_CHECKMULTISIG,
)
from test_framework.script_util import (
key_to_p2pkh_script,
key_to_p2wpkh_script,
script_to_p2sh_script,
script_to_p2wsh_script,
)
from test_framework.util import hex_str_to_bytes
Key = namedtuple('Key', ['privkey',
'pubkey',
'p2pkh_script',
'p2pkh_addr',
'p2wpkh_script',
'p2wpkh_addr',
'p2sh_p2wpkh_script',
'p2sh_p2wpkh_redeem_script',
'p2sh_p2wpkh_addr'])
Multisig = namedtuple('Multisig', ['privkeys',
'pubkeys',
'p2sh_script',
'p2sh_addr',
'redeem_script',
'p2wsh_script',
'p2wsh_addr',
'p2sh_p2wsh_script',
'p2sh_p2wsh_addr'])
def get_key(node):
"""Generate a fresh key on node
Returns a named tuple of privkey, pubkey and all address and scripts."""
addr = node.getnewaddress()
pubkey = node.getaddressinfo(addr)['pubkey']
return Key(privkey=node.dumpprivkey(addr),
pubkey=pubkey,
p2pkh_script=key_to_p2pkh_script(pubkey).hex(),
p2pkh_addr=key_to_p2pkh(pubkey),
p2wpkh_script=key_to_p2wpkh_script(pubkey).hex(),
p2wpkh_addr=key_to_p2wpkh(pubkey),
p2sh_p2wpkh_script=script_to_p2sh_script(key_to_p2wpkh_script(pubkey)).hex(),
p2sh_p2wpkh_redeem_script=key_to_p2wpkh_script(pubkey).hex(),
p2sh_p2wpkh_addr=key_to_p2sh_p2wpkh(pubkey))
def get_generate_key():
"""Generate a fresh key
Returns a named tuple of privkey, pubkey and all address and scripts."""
eckey = ECKey()
eckey.generate()
privkey = bytes_to_wif(eckey.get_bytes())
pubkey = eckey.get_pubkey().get_bytes().hex()
return Key(privkey=privkey,
pubkey=pubkey,
p2pkh_script=key_to_p2pkh_script(pubkey).hex(),
p2pkh_addr=key_to_p2pkh(pubkey),
p2wpkh_script=key_to_p2wpkh_script(pubkey).hex(),
p2wpkh_addr=key_to_p2wpkh(pubkey),
p2sh_p2wpkh_script=script_to_p2sh_script(key_to_p2wpkh_script(pubkey)).hex(),
p2sh_p2wpkh_redeem_script=key_to_p2wpkh_script(pubkey).hex(),
p2sh_p2wpkh_addr=key_to_p2sh_p2wpkh(pubkey))
def get_multisig(node):
"""Generate a fresh 2-of-3 multisig on node
Returns a named tuple of privkeys, pubkeys and all address and scripts."""
addrs = []
pubkeys = []
for _ in range(3):
addr = node.getaddressinfo(node.getnewaddress())
addrs.append(addr['address'])
pubkeys.append(addr['pubkey'])
script_code = CScript([OP_2] + [hex_str_to_bytes(pubkey) for pubkey in pubkeys] + [OP_3, OP_CHECKMULTISIG])
witness_script = script_to_p2wsh_script(script_code)
return Multisig(privkeys=[node.dumpprivkey(addr) for addr in addrs],
pubkeys=pubkeys,
p2sh_script=script_to_p2sh_script(script_code).hex(),
p2sh_addr=script_to_p2sh(script_code),
redeem_script=script_code.hex(),
p2wsh_script=witness_script.hex(),
p2wsh_addr=script_to_p2wsh(script_code),
p2sh_p2wsh_script=script_to_p2sh_script(witness_script).hex(),
p2sh_p2wsh_addr=script_to_p2sh_p2wsh(script_code))
def test_address(node, address, **kwargs):
"""Get address info for `address` and test whether the returned values are as expected."""
addr_info = node.getaddressinfo(address)
for key, value in kwargs.items():
if value is None:
if key in addr_info.keys():
raise AssertionError("key {} unexpectedly returned in getaddressinfo.".format(key))
elif addr_info[key] != value:
raise AssertionError("key {} value {} did not match expected value {}".format(key, addr_info[key], value))
def bytes_to_wif(b, compressed=True):
if compressed:
b += b'\x01'
return byte_to_base58(b, 239)
def generate_wif_key():
# Makes a WIF privkey for imports
k = ECKey()
k.generate()
return bytes_to_wif(k.get_bytes(), k.is_compressed)
| 39.640625 | 118 | 0.615294 | [
"MIT"
] | ludirium/ludirium | test/functional/test_framework/wallet_util.py | 5,074 | Python |
import pymongo
import sys
# establish a connection to the database
# note this uses the now deprecated Connection class, as we did in the lecture.
# MongoClient is the preferred way of connecting.
connection = pymongo.Connection("mongodb://localhost", safe=True)
# get a handle to the school database
db=connection.school
scores = db.scores
query = {''}
try:
doc = scores.find_one(query)
except:
print "Unexpected error:", sys.exc_info()[0]
print doc
| 23.090909 | 80 | 0.679134 | [
"Apache-2.0"
] | hemmerling/nosql-mongodb2013 | src/m101p/week02/lesson_files/hemmerling_week2_01.py | 508 | Python |
from django.contrib import admin
from .models import Arts, Comments, Tags, ArtworksTags, Stili, Umetnina, Umetnik
# Register your models here.
admin.site.register(Umetnik)
admin.site.register(Umetnina)
admin.site.register(Stili)
admin.site.register(Arts)
admin.site.register(Comments)
admin.site.register(Tags)
admin.site.register(ArtworksTags)
# admin.site.register(ArtworkLikes)
| 25.6 | 80 | 0.807292 | [
"MIT"
] | jaanos/OPB-umetnine | umetnine/artists/admin.py | 384 | Python |
# Copyright Istio Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import print_function
import collections
import os
import json
import argparse
import subprocess
import shlex
import uuid
from fortio import METRICS_START_SKIP_DURATION, METRICS_END_SKIP_DURATION
import sys
if sys.version_info.major == 2:
from commands import getoutput
else:
from subprocess import getoutput
POD = collections.namedtuple('Pod', ['name', 'namespace', 'ip', 'labels'])
def pod_info(filterstr="", namespace="twopods", multi_ok=True):
cmd = "kubectl -n {namespace} get pod {filterstr} -o json".format(
namespace=namespace, filterstr=filterstr)
op = getoutput(cmd)
o = json.loads(op)
items = o['items']
if not multi_ok and len(items) > 1:
raise Exception("more than one found " + op)
if not items:
raise Exception("no pods found with command [" + cmd + "]")
i = items[0]
return POD(i['metadata']['name'], i['metadata']['namespace'],
i['status']['podIP'], i['metadata']['labels'])
def run_command(command):
process = subprocess.Popen(shlex.split(command))
process.wait()
def run_command_sync(command):
op = getoutput(command)
return op.strip()
class Fortio:
ports = {
"http": {"direct_port": 8077, "port": 8080, "ingress": 80},
"grpc": {"direct_port": 8076, "port": 8079, "ingress": 80},
"direct_envoy": {"direct_port": 8076, "port": 8079},
}
def __init__(
self,
conn=None,
qps=None,
duration=None,
size=None,
mode="http",
mixer_mode="mixer",
mixer_cache=True,
perf_record=False,
server="fortioserver",
client="fortioclient",
additional_args=None,
filter_fn=None,
labels=None,
baseline=False,
serversidecar=False,
clientsidecar=True,
ingress=None,
mesh="istio"):
self.run_id = str(uuid.uuid4()).partition('-')[0]
self.conn = conn
self.qps = qps
self.size = size
self.duration = duration
self.mode = mode
self.ns = os.environ.get("NAMESPACE", "twopods")
# bucket resolution in seconds
self.r = "0.00005"
self.mixer_mode = mixer_mode
self.mixer_cache = mixer_cache
self.perf_record = perf_record
self.server = pod_info("-lapp=" + server, namespace=self.ns)
self.client = pod_info("-lapp=" + client, namespace=self.ns)
self.additional_args = additional_args
self.filter_fn = filter_fn
self.labels = labels
self.run_baseline = baseline
self.run_serversidecar = serversidecar
self.run_clientsidecar = clientsidecar
self.run_ingress = ingress
if mesh == "linkerd":
self.mesh = "linkerd"
elif mesh == "istio":
self.mesh = "istio"
else:
sys.exit("invalid mesh %s, must be istio or linkerd" % mesh)
def nosidecar(self, fortio_cmd):
basestr = "http://{svc}:{port}/echo?size={size}"
if self.mode == "grpc":
basestr = "-payload-size {size} {svc}:{port}"
return fortio_cmd + "_base " + basestr.format(
svc=self.server.ip, port=self.ports[self.mode]["direct_port"], size=self.size)
def serversidecar(self, fortio_cmd):
basestr = "http://{svc}:{port}/echo?size={size}"
if self.mode == "grpc":
basestr = "-payload-size {size} {svc}:{port}"
return fortio_cmd + "_serveronly " + basestr.format(
svc=self.server.ip, port=self.ports[self.mode]["port"], size=self.size)
def bothsidecar(self, fortio_cmd):
basestr = "http://{svc}:{port}/echo?size={size}"
if self.mode == "grpc":
basestr = "-payload-size {size} {svc}:{port}"
return fortio_cmd + "_both " + basestr.format(
svc=self.server.labels["app"], port=self.ports[self.mode]["port"], size=self.size)
def ingress(self, fortio_cmd):
svc = self.run_ingress
if ':' not in svc:
svc += ":{port}".format(port=self.ports[self.mode]["ingress"])
return fortio_cmd + "_ingress http://{svc}/echo?size={size}".format(
svc=svc, size=self.size)
def run(self, conn, qps, size, duration):
size = size or self.size
if duration is None:
duration = self.duration
labels = self.run_id
labels += "_qps_" + str(qps)
labels += "_c_" + str(conn)
labels += "_" + str(size)
# Mixer label
labels += "_"
labels += self.mixer_mode
if self.labels is not None:
labels += "_" + self.labels
grpc = ""
if self.mode == "grpc":
grpc = "-grpc -ping"
fortio_cmd = (
"fortio load -c {conn} -qps {qps} -t {duration}s -a -r {r} {grpc} -httpbufferkb=128 " +
"-labels {labels}").format(
conn=conn,
qps=qps,
duration=duration,
r=self.r,
grpc=grpc,
labels=labels)
if self.run_ingress:
kubectl_exec(self.client.name, self.ingress(fortio_cmd))
if self.perf_record:
run_perf(
self.mesh,
self.server.name,
labels + "_srv_ingress",
duration=40)
if self.run_serversidecar:
kubectl_exec(self.client.name, self.serversidecar(fortio_cmd))
if self.perf_record:
run_perf(
self.mesh,
self.server.name,
labels + "_srv_serveronly",
duration=40)
if self.run_clientsidecar:
kubectl_exec(self.client.name, self.bothsidecar(fortio_cmd))
if self.perf_record:
run_perf(
self.mesh,
self.server.name,
labels + "_srv_bothsidecars",
duration=40)
if self.run_baseline:
kubectl_exec(self.client.name, self.nosidecar(fortio_cmd))
PERFCMD = "/usr/lib/linux-tools/4.4.0-131-generic/perf"
PERFSH = "get_perfdata.sh"
PERFWD = "/etc/istio/proxy/"
def run_perf(mesh, pod, labels, duration=20):
filename = labels + "_perf.data"
filepath = PERFWD + filename
perfpath = PERFWD + PERFSH
# copy executable over
kubectl_cp(PERFSH, pod + ":" + perfpath, mesh + "-proxy")
kubectl_exec(
pod,
"{perf_cmd} {filename} {duration}".format(
perf_cmd=perfpath,
filename=filename,
duration=duration),
container=mesh + "-proxy")
kubectl_cp(pod + ":" + filepath + ".perf", filename + ".perf", mesh + "-proxy")
run_command_sync("../flame/flame.sh " + filename + ".perf")
def kubectl_cp(from_file, to_file, container):
namespace = os.environ.get("NAMESPACE", "twopods")
cmd = "kubectl --namespace {namespace} cp {from_file} {to_file} -c {container}".format(
namespace=namespace,
from_file=from_file,
to_file=to_file,
container=container)
print(cmd)
run_command_sync(cmd)
def kubectl_exec(pod, remote_cmd, runfn=run_command, container=None):
namespace = os.environ.get("NAMESPACE", "twopods")
c = ""
if container is not None:
c = "-c " + container
cmd = "kubectl --namespace {namespace} exec -i -t {pod} {c} -- {remote_cmd}".format(
pod=pod,
remote_cmd=remote_cmd,
c=c,
namespace=namespace)
print(cmd)
runfn(cmd)
def rc(command):
process = subprocess.Popen(command.split(), stdout=subprocess.PIPE)
while True:
output = process.stdout.readline()
if output == '' and process.poll() is not None:
break
if output:
print(output.strip() + "\n")
return process.poll()
def run(args):
min_duration = METRICS_START_SKIP_DURATION + METRICS_END_SKIP_DURATION
if args.duration <= min_duration:
print("Duration must be greater than {min_duration}".format(
min_duration=min_duration))
exit(1)
fortio = Fortio(
conn=args.conn,
qps=args.qps,
duration=args.duration,
size=args.size,
perf_record=args.perf,
labels=args.labels,
baseline=args.baseline,
serversidecar=args.serversidecar,
clientsidecar=args.clientsidecar,
ingress=args.ingress,
mode=args.mode,
mesh=args.mesh,
mixer_mode=args.mixer_mode)
for conn in args.conn:
for qps in args.qps:
fortio.run(conn=conn, qps=qps,
duration=args.duration, size=args.size)
def csv_to_int(s):
return [int(i) for i in s.split(",")]
def get_parser():
parser = argparse.ArgumentParser("Run performance test")
parser.add_argument(
"conn",
help="number of connections, comma separated list",
type=csv_to_int,)
parser.add_argument(
"qps",
help="qps, comma separated list",
type=csv_to_int,)
parser.add_argument(
"duration",
help="duration in seconds of the extract",
type=int)
parser.add_argument(
"--size",
help="size of the payload",
type=int,
default=1024)
parser.add_argument(
"--mesh",
help="istio or linkerd",
default="istio")
parser.add_argument(
"--mixer_mode",
help="run with different mixer configurations: mixer, nomixer, mixerv2",
default="mixer")
parser.add_argument(
"--client",
help="where to run the test from",
default=None)
parser.add_argument(
"--server",
help="pod ip of the server",
default=None)
parser.add_argument(
"--perf",
help="also run perf and produce flame graph",
default=False)
parser.add_argument(
"--ingress",
help="run traffic through ingress",
default=None)
parser.add_argument(
"--labels",
help="extra labels",
default=None)
parser.add_argument(
"--mode",
help="http or grpc",
default="http")
define_bool(parser, "baseline", "run baseline for all", False)
define_bool(parser, "serversidecar",
"run serversidecar-only for all", False)
define_bool(parser, "clientsidecar",
"run clientsidecar and serversidecar for all", True)
return parser
def define_bool(parser, opt, help_arg, default_val):
parser.add_argument(
"--" + opt, help=help_arg, dest=opt, action='store_true')
parser.add_argument(
"--no-" + opt, help="do not " + help_arg, dest=opt, action='store_false')
val = {opt: default_val}
parser.set_defaults(**val)
def main(argv):
args = get_parser().parse_args(argv)
print(args)
return run(args)
if __name__ == "__main__":
import sys
sys.exit(main(sys.argv[1:]))
| 30.658793 | 99 | 0.583769 | [
"Apache-2.0"
] | jwendell/tools | perf/benchmark/runner/runner.py | 11,681 | Python |
# Usage: testWordsInCorpus.py [language] {corpus file}
# If no corpus file is named, the programme will try to load a corresponding cPickle file.
#
# German corpus: /mounts/data/proj/huiming/SIGMORPHON/dewiki-20151102-pages-articles-multistream.xml
#
# This script finds words that should belong to a paradigm in the corpus and adds them (for training?).
from getEditTrees import editTreesByPos
from getEditTrees import applyOnlyTree
import sys
import pickle as cPickle
toAdd = {} # lemma to things that should be autocompleted
uniquenessCheck = {} # (lemma, form) -> word, avoiding that we add things we are unsure about
# New autocomplete. Finds union and checks if paradigms can complete each other.
# We suppose the union consists of at least 2 edit trees.
# TODO: account for Umlaute.
# Returns a dictinary lemma -> (et, tags) with things to add to the original one.
# TODO: irgendwas stimmt hier nicht. korrigiere es
def autoComplete(lemma1, etTag1, lemma2, etTag2, corpusWords):
etAndTagToAdd = set()
notFound = 0
allRight1 = True
allRight2 = True
for (et, form) in etTag1.difference(etTag2):
result = applyOnlyTree(lemma2, et)
if result == '#error#':
allRight = False
break
if result not in corpusWords or corpusWords[result] <=3: # orig is 3
notFound += 1
if notFound == 2:
allRight = False
break
else:
etAndTagToAdd.add((et, form))
if allRight and etAndTagToAdd:
if lemma2 not in toAdd:
toAdd[lemma2] = set()
toAdd[lemma2] = toAdd[lemma2].union(etAndTagToAdd)
for (et, form) in etAndTagToAdd:
if (lemma2, form) not in uniquenessCheck:
uniquenessCheck[(lemma2, form)] = set()
else:
if applyOnlyTree(lemma2,et) not in uniquenessCheck[(lemma2, form)]:
print("yeay")
uniquenessCheck[(lemma2, form)].add(applyOnlyTree(lemma2, et))
# Lemma 1 has more ETs than lemma 2.
# Returns a dictinary lemma -> (et, tags) with things to add to the original one.
def autoComplete2(lemma1, etTag1, lemma2, etTag2, corpusWords):
etAndTagToAdd = set()
notFound = 0
allRight = True
for (et, form) in etTag1.difference(etTag2):
result = applyOnlyTree(lemma2, et)
if result == '#error#':
allRight = False
break
if result not in corpusWords or corpusWords[result] <=3: # orig is 3
notFound += 1
if notFound == 2:
allRight = False
break
else:
etAndTagToAdd.add((et, form))
if allRight and etAndTagToAdd:
if lemma2 not in toAdd:
toAdd[lemma2] = set()
toAdd[lemma2] = toAdd[lemma2].union(etAndTagToAdd)
for (et, form) in etAndTagToAdd:
if (lemma2, form) not in uniquenessCheck:
uniquenessCheck[(lemma2, form)] = set()
uniquenessCheck[(lemma2, form)].add(applyOnlyTree(lemma2, et))
# Test if a group of (edit tree, tag) combinations for a lemma is subset of the one for another lemma.
# If yes, try if the missing edit trees are applicable and if the corresponding word appears in the corpus.
def getAdditionalWords(lemmaToEtAndTag, corpusWords):
isTrue = 0
isFalse = 0
for lemma1, etTag1 in lemmaToEtAndTag.items():
for lemma2, etTag2 in lemmaToEtAndTag.items():
if len(etTag1) <= 1 or len(etTag2) <= 1: # for now, don't complete things with 0 or only 1 entry. We are just not sure enough.
isFalse += 1
continue
maybeSame = False
if len(etTag1) > len(etTag2)+2:
if len(etTag1) >= 3 and len(etTag2.union(etTag1)) > 1 and etTag2.issubset(etTag1):
maybeSame = True
autoComplete(lemma1, etTag1, lemma2, etTag2, corpusWords)
isTrue += 1
else:
isFalse += 1
elif len(etTag2) > len(etTag1)+2:
if len(etTag2) >= 3 and len(etTag2.union(etTag1)) > 1 and etTag1.issubset(etTag2):
maybeSame = True
autoComplete(lemma2, etTag2, lemma1, etTag1, corpusWords)
isTrue += 1
else:
isFalse += 1
#print(str(len(toAdd)) + ' words have been added.')
#print("Is subset: " + str(isTrue))
#print("No subset: " + str(isFalse))
#sys.exit(0)
noWordsToAdd = 0
for lemma, aSet in toAdd.items():
noWordsToAdd += len(aSet)
'''
for (lemma, form), word in uniquenessCheck.items():
if len(word) > 1:
print(word)
sys.exit(0)
'''
return noWordsToAdd
def announce(*objs):
print("# ", *objs, file = sys.stderr)
if __name__ == "__main__":
lang = sys.argv[1]
if len(sys.argv) == 2:
usePickle = True
else:
usePickle = False
posToEt, lemmaToEtAndTag = editTreesByPos(lang)
for lemma, aSet in lemmaToEtAndTag.items():
for (et, form) in aSet:
if (lemma, form) not in uniquenessCheck:
uniquenessCheck[(lemma, form)] = set()
uniquenessCheck[(lemma, form)].add(applyOnlyTree(lemma, et))
#print(applyOnlyTree(lemma, et))
#sys.exit(0)
if not usePickle:
# Read the bonus corpus.
announce('Start reading corpus...')
corpusWords = {} # word to its frequency
with open(sys.argv[2], 'r') as corpus_file:
for line in corpus_file:
#tokens = tokenize.word_tokenize(line.strip())
tokens = line.strip().split(' ')
for token in tokens:
if token not in corpusWords:
corpusWords[token] = 0
corpusWords[token] += 1
announce('Done reading corpus.')
# Store the dictionary to a binary file.
print('Store the dictionary with the corpus words to a binary file...')
save_file = open('/mounts/data/proj/huiming/SIGMORPHON/corpusWords_' + lang, 'wb')
cPickle.dump(corpusWords, save_file, -1)
save_file.close()
print('Done.')
else:
# Load the corpusWords dictionary.
announce('Load the words with cPickle...')
vocListFile = open('/mounts/data/proj/huiming/SIGMORPHON/corpusWords_' + lang, 'rb')
corpusWords = cPickle.load(vocListFile)
vocListFile.close()
announce('Words loaded.')
lastNumber = 0
noWordsToAdd = 1
while noWordsToAdd > lastNumber:
lastNumber = noWordsToAdd
noWordsToAdd = getAdditionalWords(lemmaToEtAndTag, corpusWords)
for lemma, aSet in lemmaToEtAndTag.items():
if lemma in toAdd:
lemmaToEtAndTag[lemma] = lemmaToEtAndTag[lemma].union(toAdd[lemma])
announce('Number word to add: ' + str(noWordsToAdd))
# The union did not work well for some reason. Therefore, use toAdd directly.
additionalWordsCounter = 0
with open('/mounts/Users/cisintern/huiming/SIGMORPHON/Code/data/' + lang + '-bigger-task1-train', 'w') as out_file:
with open('/mounts/Users/cisintern/huiming/SIGMORPHON/Code/data/' + lang + '-task1-train', 'r') as original_file:
for line in original_file:
out_file.write(line)
for lemma, etAndTagSet in toAdd.items():
for (et, form) in etAndTagSet:
if len(uniquenessCheck[(lemma, form)]) > 1:
continue
out_file.write(lemma + '\t' + form + '\t' + applyOnlyTree(lemma, et) + '\n')
additionalWordsCounter += 1
print(str(additionalWordsCounter) + ' words have been added.')
| 34.570048 | 132 | 0.654556 | [
"MIT"
] | oncebasun/seq2seq-theano | MyAlgorithm/addWordsToParadigms_old.py | 7,156 | Python |
#!/usr/bin/env python2
from setuptools import setup
from setuptools import find_packages
setup(
name="rover",
version="0.1",
description="Algorithm for risk and sensor quality aware sensor" +
"coverage for quadrotors",
author="Alex Wallar",
author_email="[email protected]",
packages=find_packages(),
install_requires=[
"numpy",
"scipy"
],
data_files=[
(
'config',
['configs/config.json'],
)
]
)
| 20.04 | 70 | 0.59481 | [
"Apache-2.0"
] | wallarelvo/rover | setup.py | 501 | Python |
#!/usr/bin/python
#coding:utf-8
import time
import json
import requests
from selenium import webdriver
filename = 'a.csv'
url = 'http://www.icourse163.org/university/view/all.htm#/'
headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0', 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8'}
# with open(filename, 'w+') as file:
# file.write("大学,课程,课程时长,课程负载,内容类型,课程分类\n")
file = open(filename, 'w+')
print("大学,课程,课程时长,课程负载,内容类型,课程分类")
file.write("大学,课程,课程时长,课程负载,内容类型,课程分类\n")
browser = webdriver.PhantomJS()
browser2 = webdriver.PhantomJS()
browser3 = webdriver.PhantomJS()
browser.get(url)
# 大学
university = browser.find_elements_by_class_name("u-usity")
for i in university:
university_url = i.get_attribute("href")
university_name = i.find_element_by_id("").get_attribute("alt")
browser2.get(university_url)
# 课程
course = browser2.find_elements_by_class_name("g-cell1")
for j in course:
course_url = "http://www.icourse163.org" + j.get_attribute("data-href")
course_name = j.find_element_by_class_name("card").find_element_by_class_name("f-f0").text
browser3.get(course_url)
# 课程信息
course_text = browser3.find_elements_by_class_name("block")
try:
k0 = course_text[0].find_element_by_class_name("t2").text
k1 = course_text[1].find_element_by_class_name("t2").text
k2 = course_text[2].find_element_by_class_name("t2").text
k3 = course_text[3].find_element_by_class_name("t2").text
except Exception as e:
k3 = k2
k2 = k1
k1 = None
K0 = None
finally:
print("%s,%s,%s,%s,%s,%s" % (university_name,course_name,k0,k1,k2,k3))
file.write("%s,%s,%s,%s,%s,%s\n" % (university_name,course_name,k0,k1,k2,k3))
# with open(filename, 'a+') as file:
# file.write("%s,%s,%s,%s,%s,%s\n" % (university_name,course_name,k0,k1,k2,k3))
browser3.close()
browser2.close()
browser.close()
| 32.307692 | 183 | 0.645714 | [
"MIT"
] | yeonzi/163course_spider | spider.py | 2,236 | Python |
#!/usr/bin/env python
# This example uses Uvicorn package that must be installed. However, it can be
# replaced with any other ASGI-compliant server.
#
# NOTE: Python 3.6 requires aiocontextvars package to be installed.
#
# Run: python app_global_request.py
import rollbar
import uvicorn
from rollbar.contrib.starlette import LoggerMiddleware
from starlette.applications import Starlette
from starlette.responses import JSONResponse
# Integrate Rollbar with Starlette application
app = Starlette()
app.add_middleware(LoggerMiddleware) # should be added as the last middleware
async def get_user_agent():
# Global access to the current request object
request = rollbar.get_request()
user_agent = request.headers['User-Agent']
return user_agent
# $ curl -i http://localhost:8888
@app.route('/')
async def root(request):
user_agent = await get_user_agent()
return JSONResponse({'user-agent': user_agent})
if __name__ == '__main__':
uvicorn.run(app, host='localhost', port=8888)
| 26.025641 | 78 | 0.759606 | [
"MIT"
] | Apep8/pyrollbar | rollbar/examples/starlette/app_global_request.py | 1,015 | Python |
# Configuration file for jupyter-notebook.
#------------------------------------------------------------------------------
# Configurable configuration
#------------------------------------------------------------------------------
#------------------------------------------------------------------------------
# SingletonConfigurable configuration
#------------------------------------------------------------------------------
# A configurable that only allows one instance.
#
# This class is for classes that should only have one instance of itself or
# *any* subclass. To create and retrieve such a class use the
# :meth:`SingletonConfigurable.instance` method.
#------------------------------------------------------------------------------
# Application configuration
#------------------------------------------------------------------------------
# This is an application.
# The date format used by logging formatters for %(asctime)s
# c.Application.log_datefmt = '%Y-%m-%d %H:%M:%S'
# The Logging format template
# c.Application.log_format = '[%(name)s]%(highlevel)s %(message)s'
# Set the log level by value or name.
# c.Application.log_level = 30
#------------------------------------------------------------------------------
# JupyterApp configuration
#------------------------------------------------------------------------------
# Base class for Jupyter applications
# Answer yes to any prompts.
c.JupyterApp.answer_yes = True
# Full path of a config file.
# c.JupyterApp.config_file = u''
# Generate default config file.
# c.JupyterApp.generate_config = False
# Specify a config file to load.
# c.JupyterApp.config_file_name = u''
#------------------------------------------------------------------------------
# NotebookApp configuration
#------------------------------------------------------------------------------
# The number of additional ports to try if the specified port is not available.
c.NotebookApp.port_retries = 0
# Extra variables to supply to jinja templates when rendering.
# c.NotebookApp.jinja_template_vars = traitlets.Undefined
# The url for MathJax.js.
# c.NotebookApp.mathjax_url = ''
# Supply extra arguments that will be passed to Jinja environment.
# c.NotebookApp.jinja_environment_options = traitlets.Undefined
# The IP address the notebook server will listen on.
c.NotebookApp.ip = '*'
# DEPRECATED use base_url
# c.NotebookApp.base_project_url = '/'
# Python modules to load as notebook server extensions. This is an experimental
# API, and may change in future releases.
# c.NotebookApp.server_extensions = traitlets.Undefined
# Note: These extensions require the ~/.jupyter path to exist otherwise, errors will occur on startup
c.NotebookApp.server_extensions=['ipyparallel.nbextension']
# The random bytes used to secure cookies. By default this is a new random
# number every time you start the Notebook. Set it to a value in a config file
# to enable logins to persist across server sessions.
#
# Note: Cookie secrets should be kept private, do not share config files with
# cookie_secret stored in plaintext (you can read the value from a file).
# c.NotebookApp.cookie_secret = ''
# The default URL to redirect to from `/`
# c.NotebookApp.default_url = '/tree'
# The port the notebook server will listen on.
c.NotebookApp.port = 8754
# The kernel spec manager class to use. Should be a subclass of
# `jupyter_client.kernelspec.KernelSpecManager`.
#
# The Api of KernelSpecManager is provisional and might change without warning
# between this version of IPython and the next stable one.
# c.NotebookApp.kernel_spec_manager_class = <class 'jupyter_client.kernelspec.KernelSpecManager'>
# Set the Access-Control-Allow-Origin header
#
# Use '*' to allow any origin to access your server.
#
# Takes precedence over allow_origin_pat.
c.NotebookApp.allow_origin = '*'
# The notebook manager class to use.
# c.NotebookApp.contents_manager_class = <class 'notebook.services.contents.filemanager.FileContentsManager'>
# Use a regular expression for the Access-Control-Allow-Origin header
#
# Requests from an origin matching the expression will get replies with:
#
# Access-Control-Allow-Origin: origin
#
# where `origin` is the origin of the request.
#
# Ignored if allow_origin is set.
# c.NotebookApp.allow_origin_pat = ''
# The full path to an SSL/TLS certificate file.
# c.NotebookApp.certfile = u''
# The logout handler class to use.
# c.NotebookApp.logout_handler_class = <class 'notebook.auth.logout.LogoutHandler'>
# The base URL for the notebook server.
#
# Leading and trailing slashes can be omitted, and will automatically be added.
c.NotebookApp.base_url = '/'
# The session manager class to use.
# c.NotebookApp.session_manager_class = <class 'notebook.services.sessions.sessionmanager.SessionManager'>
# Supply overrides for the tornado.web.Application that the IPython notebook
# uses.
# c.NotebookApp.tornado_settings = traitlets.Undefined
# The directory to use for notebooks and kernels.
c.NotebookApp.notebook_dir = u'/root/pipeline/myapps/jupyter/'
# The kernel manager class to use.
# c.NotebookApp.kernel_manager_class = <class 'notebook.services.kernels.kernelmanager.MappingKernelManager'>
# The file where the cookie secret is stored.
# c.NotebookApp.cookie_secret_file = u''
# Supply SSL options for the tornado HTTPServer. See the tornado docs for
# details.
# c.NotebookApp.ssl_options = traitlets.Undefined
#
# c.NotebookApp.file_to_run = ''
# DISABLED: use %pylab or %matplotlib in the notebook to enable matplotlib.
# c.NotebookApp.pylab = 'disabled'
# Whether to enable MathJax for typesetting math/TeX
#
# MathJax is the javascript library IPython uses to render math/LaTeX. It is
# very large, so you may want to disable it if you have a slow internet
# connection, or for offline use of the notebook.
#
# When disabled, equations etc. will appear as their untransformed TeX source.
# c.NotebookApp.enable_mathjax = True
# Reraise exceptions encountered loading server extensions?
# c.NotebookApp.reraise_server_extension_failures = False
# The base URL for websockets, if it differs from the HTTP server (hint: it
# almost certainly doesn't).
#
# Should be in the form of an HTTP origin: ws[s]://hostname[:port]
# c.NotebookApp.websocket_url = ''
# Whether to open in a browser after starting. The specific browser used is
# platform dependent and determined by the python standard library `webbrowser`
# module, unless it is overridden using the --browser (NotebookApp.browser)
# configuration option.
c.NotebookApp.open_browser = False
# Hashed password to use for web authentication.
#
# To generate, type in a python/IPython shell:
#
# from notebook.auth import passwd; passwd()
#
# The string should be of the form type:salt:hashed-password.
# c.NotebookApp.password = u''
# extra paths to look for Javascript notebook extensions
# c.NotebookApp.extra_nbextensions_path = traitlets.Undefined
# Set the Access-Control-Allow-Credentials: true header
# c.NotebookApp.allow_credentials = False
# Extra paths to search for serving static files.
#
# This allows adding javascript/css to be available from the notebook server
# machine, or overriding individual files in the IPython
# c.NotebookApp.extra_static_paths = traitlets.Undefined
# The login handler class to use.
# c.NotebookApp.login_handler_class = <class 'notebook.auth.login.LoginHandler'>
# Whether to trust or not X-Scheme/X-Forwarded-Proto and X-Real-Ip/X-Forwarded-
# For headerssent by the upstream reverse proxy. Necessary if the proxy handles
# SSL
# c.NotebookApp.trust_xheaders = False
# Extra paths to search for serving jinja templates.
#
# Can be used to override templates from notebook.templates.
# c.NotebookApp.extra_template_paths = traitlets.Undefined
# The config manager class to use
# c.NotebookApp.config_manager_class = <class 'notebook.services.config.manager.ConfigManager'>
# The full path to a private key file for usage with SSL/TLS.
# c.NotebookApp.keyfile = u''
# DEPRECATED, use tornado_settings
# c.NotebookApp.webapp_settings = traitlets.Undefined
# Specify what command to use to invoke a web browser when opening the notebook.
# If not specified, the default browser will be determined by the `webbrowser`
# standard library module, which allows setting of the BROWSER environment
# variable to override it.
# c.NotebookApp.browser = u''
#------------------------------------------------------------------------------
# LoggingConfigurable configuration
#------------------------------------------------------------------------------
# A parent class for Configurables that log.
#
# Subclasses have a log trait, and the default behavior is to get the logger
# from the currently running Application.
#------------------------------------------------------------------------------
# ConnectionFileMixin configuration
#------------------------------------------------------------------------------
# Mixin for configurable classes that work with connection files
# set the stdin (ROUTER) port [default: random]
# c.ConnectionFileMixin.stdin_port = 0
# Set the kernel's IP address [default localhost]. If the IP address is
# something other than localhost, then Consoles on other machines will be able
# to connect to the Kernel, so be careful!
# c.ConnectionFileMixin.ip = u''
# JSON file in which to store connection info [default: kernel-<pid>.json]
#
# This file will contain the IP, ports, and authentication key needed to connect
# clients to this kernel. By default, this file will be created in the security
# dir of the current profile, but can be specified by absolute path.
# c.ConnectionFileMixin.connection_file = ''
# set the control (ROUTER) port [default: random]
# c.ConnectionFileMixin.control_port = 0
# set the heartbeat port [default: random]
# c.ConnectionFileMixin.hb_port = 0
# set the shell (ROUTER) port [default: random]
# c.ConnectionFileMixin.shell_port = 0
#
# c.ConnectionFileMixin.transport = 'tcp'
# set the iopub (PUB) port [default: random]
# c.ConnectionFileMixin.iopub_port = 0
#------------------------------------------------------------------------------
# KernelManager configuration
#------------------------------------------------------------------------------
# Manages a single kernel in a subprocess on this host.
#
# This version starts kernels with Popen.
# DEPRECATED: Use kernel_name instead.
#
# The Popen Command to launch the kernel. Override this if you have a custom
# kernel. If kernel_cmd is specified in a configuration file, Jupyter does not
# pass any arguments to the kernel, because it cannot make any assumptions about
# the arguments that the kernel understands. In particular, this means that the
# kernel does not receive the option --debug if it given on the Jupyter command
# line.
# c.KernelManager.kernel_cmd = traitlets.Undefined
# Should we autorestart the kernel if it dies.
# c.KernelManager.autorestart = False
#------------------------------------------------------------------------------
# Session configuration
#------------------------------------------------------------------------------
# Object for handling serialization and sending of messages.
#
# The Session object handles building messages and sending them with ZMQ sockets
# or ZMQStream objects. Objects can communicate with each other over the
# network via Session objects, and only need to work with the dict-based IPython
# message spec. The Session will handle serialization/deserialization, security,
# and metadata.
#
# Sessions support configurable serialization via packer/unpacker traits, and
# signing with HMAC digests via the key/keyfile traits.
#
# Parameters ----------
#
# debug : bool
# whether to trigger extra debugging statements
# packer/unpacker : str : 'json', 'pickle' or import_string
# importstrings for methods to serialize message parts. If just
# 'json' or 'pickle', predefined JSON and pickle packers will be used.
# Otherwise, the entire importstring must be used.
#
# The functions must accept at least valid JSON input, and output *bytes*.
#
# For example, to use msgpack:
# packer = 'msgpack.packb', unpacker='msgpack.unpackb'
# pack/unpack : callables
# You can also set the pack/unpack callables for serialization directly.
# session : bytes
# the ID of this Session object. The default is to generate a new UUID.
# username : unicode
# username added to message headers. The default is to ask the OS.
# key : bytes
# The key used to initialize an HMAC signature. If unset, messages
# will not be signed or checked.
# keyfile : filepath
# The file containing a key. If this is set, `key` will be initialized
# to the contents of the file.
# Username for the Session. Default is your system username.
# c.Session.username = u'username'
# Threshold (in bytes) beyond which a buffer should be sent without copying.
# c.Session.copy_threshold = 65536
# The name of the packer for serializing messages. Should be one of 'json',
# 'pickle', or an import name for a custom callable serializer.
# c.Session.packer = 'json'
# Metadata dictionary, which serves as the default top-level metadata dict for
# each message.
# c.Session.metadata = traitlets.Undefined
# The maximum number of digests to remember.
#
# The digest history will be culled when it exceeds this value.
# c.Session.digest_history_size = 65536
# The UUID identifying this session.
# c.Session.session = u''
# The digest scheme used to construct the message signatures. Must have the form
# 'hmac-HASH'.
# c.Session.signature_scheme = 'hmac-sha256'
# execution key, for signing messages.
# c.Session.key = ''
# Debug output in the Session
# c.Session.debug = False
# The name of the unpacker for unserializing messages. Only used with custom
# functions for `packer`.
# c.Session.unpacker = 'json'
# path to file containing execution key.
# c.Session.keyfile = ''
# Threshold (in bytes) beyond which an object's buffer should be extracted to
# avoid pickling.
# c.Session.buffer_threshold = 1024
# The maximum number of items for a container to be introspected for custom
# serialization. Containers larger than this are pickled outright.
# c.Session.item_threshold = 64
#------------------------------------------------------------------------------
# MultiKernelManager configuration
#------------------------------------------------------------------------------
# A class for managing multiple kernels.
# The name of the default kernel to start
# c.MultiKernelManager.default_kernel_name = 'python2'
# The kernel manager class. This is configurable to allow subclassing of the
# KernelManager for customized behavior.
# c.MultiKernelManager.kernel_manager_class = 'jupyter_client.ioloop.IOLoopKernelManager'
#------------------------------------------------------------------------------
# MappingKernelManager configuration
#------------------------------------------------------------------------------
# A KernelManager that handles notebook mapping and HTTP error handling
#
# c.MappingKernelManager.root_dir = u''
#------------------------------------------------------------------------------
# ContentsManager configuration
#------------------------------------------------------------------------------
# Base class for serving files and directories.
#
# This serves any text or binary file, as well as directories, with special
# handling for JSON notebook documents.
#
# Most APIs take a path argument, which is always an API-style unicode path, and
# always refers to a directory.
#
# - unicode, not url-escaped
# - '/'-separated
# - leading and trailing '/' will be stripped
# - if unspecified, path defaults to '',
# indicating the root path.
# The base name used when creating untitled files.
# c.ContentsManager.untitled_file = 'untitled'
# Python callable or importstring thereof
#
# To be called on a contents model prior to save.
#
# This can be used to process the structure, such as removing notebook outputs
# or other side effects that should not be saved.
#
# It will be called as (all arguments passed by keyword)::
#
# hook(path=path, model=model, contents_manager=self)
#
# - model: the model to be saved. Includes file contents.
# Modifying this dict will affect the file that is stored.
# - path: the API path of the save destination
# - contents_manager: this ContentsManager instance
# c.ContentsManager.pre_save_hook = None
#
# c.ContentsManager.checkpoints_class = <class 'notebook.services.contents.checkpoints.Checkpoints'>
# Glob patterns to hide in file and directory listings.
# c.ContentsManager.hide_globs = traitlets.Undefined
# The base name used when creating untitled notebooks.
# c.ContentsManager.untitled_notebook = 'Untitled'
# The base name used when creating untitled directories.
# c.ContentsManager.untitled_directory = 'Untitled Folder'
#
# c.ContentsManager.checkpoints = traitlets.Undefined
#
# c.ContentsManager.checkpoints_kwargs = traitlets.Undefined
#------------------------------------------------------------------------------
# FileContentsManager configuration
#------------------------------------------------------------------------------
# DEPRECATED, use post_save_hook
# c.FileContentsManager.save_script = False
#
# c.FileContentsManager.root_dir = u''
# Python callable or importstring thereof
#
# to be called on the path of a file just saved.
#
# This can be used to process the file on disk, such as converting the notebook
# to a script or HTML via nbconvert.
#
# It will be called as (all arguments passed by keyword)::
#
# hook(os_path=os_path, model=model, contents_manager=instance)
#
# - path: the filesystem path to the file just written - model: the model
# representing the file - contents_manager: this ContentsManager instance
# c.FileContentsManager.post_save_hook = None
#------------------------------------------------------------------------------
# NotebookNotary configuration
#------------------------------------------------------------------------------
# A class for computing and verifying notebook signatures.
# The number of notebook signatures to cache. When the number of signatures
# exceeds this value, the oldest 25% of signatures will be culled.
# c.NotebookNotary.cache_size = 65535
# The secret key with which notebooks are signed.
# c.NotebookNotary.secret = ''
# The sqlite file in which to store notebook signatures. By default, this will
# be in your Jupyter runtime directory. You can set it to ':memory:' to disable
# sqlite writing to the filesystem.
# c.NotebookNotary.db_file = u''
# The hashing algorithm used to sign notebooks.
# c.NotebookNotary.algorithm = 'sha256'
# The file where the secret key is stored.
# c.NotebookNotary.secret_file = u''
#------------------------------------------------------------------------------
# KernelSpecManager configuration
#------------------------------------------------------------------------------
# Whitelist of allowed kernel names.
#
# By default, all installed kernels are allowed.
# c.KernelSpecManager.whitelist = traitlets.Undefined
| 36.942197 | 109 | 0.668388 | [
"Apache-2.0"
] | TrinathY/pipeline | config/jupyter/jupyter_notebook_config.py | 19,173 | Python |
import asyncio
import logging
import os
import shutil
import warnings
from types import TracebackType
from typing import Any, Coroutine, Dict, List, Optional, Text, Type, TypeVar
import rasa.core.utils
import rasa.utils.io
from rasa.constants import (
DEFAULT_LOG_LEVEL_LIBRARIES,
ENV_LOG_LEVEL_LIBRARIES,
)
from rasa.shared.constants import DEFAULT_LOG_LEVEL, ENV_LOG_LEVEL
import rasa.shared.utils.io
logger = logging.getLogger(__name__)
T = TypeVar("T")
class TempDirectoryPath(str):
"""Represents a path to an temporary directory. When used as a context
manager, it erases the contents of the directory on exit.
"""
def __enter__(self) -> "TempDirectoryPath":
return self
def __exit__(
self,
_exc: Optional[Type[BaseException]],
_value: Optional[Exception],
_tb: Optional[TracebackType],
) -> bool:
if os.path.exists(self):
shutil.rmtree(self)
def read_global_config(path: Text) -> Dict[Text, Any]:
"""Read global Rasa configuration.
Args:
path: Path to the configuration
Returns:
The global configuration
"""
# noinspection PyBroadException
try:
return rasa.shared.utils.io.read_config_file(path)
except Exception:
# if things go south we pretend there is no config
return {}
def set_log_level(log_level: Optional[int] = None):
"""Set log level of Rasa and Tensorflow either to the provided log level or
to the log level specified in the environment variable 'LOG_LEVEL'. If none is set
a default log level will be used."""
if not log_level:
log_level = os.environ.get(ENV_LOG_LEVEL, DEFAULT_LOG_LEVEL)
log_level = logging.getLevelName(log_level)
logging.getLogger("rasa").setLevel(log_level)
update_tensorflow_log_level()
update_asyncio_log_level()
update_apscheduler_log_level()
update_socketio_log_level()
os.environ[ENV_LOG_LEVEL] = logging.getLevelName(log_level)
def update_apscheduler_log_level() -> None:
log_level = os.environ.get(ENV_LOG_LEVEL_LIBRARIES, DEFAULT_LOG_LEVEL_LIBRARIES)
apscheduler_loggers = [
"apscheduler",
"apscheduler.scheduler",
"apscheduler.executors",
"apscheduler.executors.default",
]
for logger_name in apscheduler_loggers:
logging.getLogger(logger_name).setLevel(log_level)
logging.getLogger(logger_name).propagate = False
def update_socketio_log_level() -> None:
log_level = os.environ.get(ENV_LOG_LEVEL_LIBRARIES, DEFAULT_LOG_LEVEL_LIBRARIES)
socketio_loggers = ["websockets.protocol", "engineio.server", "socketio.server"]
for logger_name in socketio_loggers:
logging.getLogger(logger_name).setLevel(log_level)
logging.getLogger(logger_name).propagate = False
def update_tensorflow_log_level() -> None:
"""Set the log level of Tensorflow to the log level specified in the environment
variable 'LOG_LEVEL_LIBRARIES'."""
# Disables libvinfer, tensorRT, cuda, AVX2 and FMA warnings (CPU support). This variable needs to be set before the
# first import since some warnings are raised on the first import.
os.environ["TF_CPP_MIN_LOG_LEVEL"] = "2"
import tensorflow as tf
log_level = os.environ.get(ENV_LOG_LEVEL_LIBRARIES, DEFAULT_LOG_LEVEL_LIBRARIES)
if log_level == "DEBUG":
tf_log_level = tf.compat.v1.logging.DEBUG
elif log_level == "INFO":
tf_log_level = tf.compat.v1.logging.INFO
elif log_level == "WARNING":
tf_log_level = tf.compat.v1.logging.WARN
else:
tf_log_level = tf.compat.v1.logging.ERROR
tf.compat.v1.logging.set_verbosity(tf_log_level)
logging.getLogger("tensorflow").propagate = False
def update_sanic_log_level(log_file: Optional[Text] = None):
"""Set the log level of sanic loggers to the log level specified in the environment
variable 'LOG_LEVEL_LIBRARIES'."""
from sanic.log import logger, error_logger, access_logger
log_level = os.environ.get(ENV_LOG_LEVEL_LIBRARIES, DEFAULT_LOG_LEVEL_LIBRARIES)
logger.setLevel(log_level)
error_logger.setLevel(log_level)
access_logger.setLevel(log_level)
logger.propagate = False
error_logger.propagate = False
access_logger.propagate = False
if log_file is not None:
formatter = logging.Formatter("%(asctime)s [%(levelname)-5.5s] %(message)s")
file_handler = logging.FileHandler(log_file)
file_handler.setFormatter(formatter)
logger.addHandler(file_handler)
error_logger.addHandler(file_handler)
access_logger.addHandler(file_handler)
def update_asyncio_log_level() -> None:
"""Set the log level of asyncio to the log level specified in the environment
variable 'LOG_LEVEL_LIBRARIES'."""
log_level = os.environ.get(ENV_LOG_LEVEL_LIBRARIES, DEFAULT_LOG_LEVEL_LIBRARIES)
logging.getLogger("asyncio").setLevel(log_level)
def set_log_and_warnings_filters() -> None:
"""
Set log filters on the root logger, and duplicate filters for warnings.
Filters only propagate on handlers, not loggers.
"""
for handler in logging.getLogger().handlers:
handler.addFilter(RepeatedLogFilter())
warnings.filterwarnings("once", category=UserWarning)
def obtain_verbosity() -> int:
"""Returns a verbosity level according to the set log level."""
log_level = os.environ.get(ENV_LOG_LEVEL, DEFAULT_LOG_LEVEL)
verbosity = 0
if log_level == "DEBUG":
verbosity = 2
if log_level == "INFO":
verbosity = 1
return verbosity
def sort_list_of_dicts_by_first_key(dicts: List[Dict]) -> List[Dict]:
"""Sorts a list of dictionaries by their first key."""
return sorted(dicts, key=lambda d: list(d.keys())[0])
def write_global_config_value(name: Text, value: Any) -> None:
"""Read global Rasa configuration."""
# need to use `rasa.constants.GLOBAL_USER_CONFIG_PATH` to allow patching
# in tests
config_path = rasa.constants.GLOBAL_USER_CONFIG_PATH
try:
os.makedirs(os.path.dirname(config_path), exist_ok=True)
c = read_global_config(config_path)
c[name] = value
rasa.core.utils.dump_obj_as_yaml_to_file(
rasa.constants.GLOBAL_USER_CONFIG_PATH, c
)
except Exception as e:
logger.warning(f"Failed to write global config. Error: {e}. Skipping.")
def read_global_config_value(name: Text, unavailable_ok: bool = True) -> Any:
"""Read a value from the global Rasa configuration."""
def not_found():
if unavailable_ok:
return None
else:
raise ValueError(f"Configuration '{name}' key not found.")
# need to use `rasa.constants.GLOBAL_USER_CONFIG_PATH` to allow patching
# in tests
config_path = rasa.constants.GLOBAL_USER_CONFIG_PATH
if not os.path.exists(config_path):
return not_found()
c = read_global_config(config_path)
if name in c:
return c[name]
else:
return not_found()
def update_existing_keys(
original: Dict[Any, Any], updates: Dict[Any, Any]
) -> Dict[Any, Any]:
"""Iterate through all the updates and update a value in the original dictionary.
If the updates contain a key that is not present in the original dict, it will
be ignored."""
updated = original.copy()
for k, v in updates.items():
if k in updated:
updated[k] = v
return updated
class RepeatedLogFilter(logging.Filter):
"""Filter repeated log records."""
last_log = None
def filter(self, record):
current_log = (
record.levelno,
record.pathname,
record.lineno,
record.msg,
record.args,
)
if current_log != self.last_log:
self.last_log = current_log
return True
return False
def run_in_loop(
f: Coroutine[Any, Any, T], loop: Optional[asyncio.AbstractEventLoop] = None
) -> T:
"""Execute the awaitable in the passed loop.
If no loop is passed, the currently existing one is used or a new one is created
if no loop has been started in the current context.
After the awaitable is finished, all remaining tasks on the loop will be
awaited as well (background tasks).
WARNING: don't use this if there are never ending background tasks scheduled.
in this case, this function will never return.
Args:
f: function to execute
loop: loop to use for the execution
Returns:
return value from the function
"""
if loop is None:
try:
loop = asyncio.get_event_loop()
except RuntimeError:
loop = asyncio.new_event_loop()
asyncio.set_event_loop(loop)
result = loop.run_until_complete(f)
# Let's also finish all running tasks:
pending = asyncio.Task.all_tasks()
loop.run_until_complete(asyncio.gather(*pending))
return result
| 30.01 | 119 | 0.690103 | [
"Apache-2.0"
] | karen-white/rasa | rasa/utils/common.py | 9,003 | Python |
from io import BytesIO
from gtts import gTTS
from PIL import Image
from vkbottle import AudioUploader, Bot, DocUploader, Message, PhotoUploader
bot = Bot("token")
photo_uploader = PhotoUploader(bot.api, generate_attachment_strings=True)
doc_uploader = DocUploader(bot.api, generate_attachment_strings=True)
audio_uploader = AudioUploader(bot.api, generate_attachment_strings=True)
@bot.on.message_handler(text="photo_from_bytes", lower=True)
async def photo_from_bytes(ans: Message):
image = Image.new("RGB", (320, 320), (0, 0, 0))
fp = BytesIO()
image.save(fp, "RGB")
setattr(fp, "name", "image.png")
photo = await photo_uploader.upload_message_photo(fp)
await ans(attachment=photo)
@bot.on.message_handler(text="doc_from_file", lower=True)
async def photo_from_bytes(ans: Message):
image = Image.new("RGB", (320, 320), (0, 0, 0))
image.save("image.png", "RGB")
photo = await doc_uploader.upload_doc_to_message("image.png", ans.peer_id)
await ans(attachment=photo)
@bot.on.message_handler(text="audio_message")
async def audio(ans: Message):
tts = gTTS(text="бокале монада", lang="ru")
fp = BytesIO()
tts.write_to_fp(fp)
audio_message = await audio_uploader.upload_audio_message(fp, ans.peer_id)
await ans(attachment=audio_message)
if __name__ == "__main__":
bot.run_polling()
| 31.465116 | 78 | 0.731707 | [
"MIT"
] | MooFreak/vkbottle | examples/uploaders.py | 1,365 | Python |
"""
ECB没有偏移量
"""
from Crypto.Cipher import AES
from binascii import b2a_hex, a2b_hex
from utils import DES_decrypt, DES_encrypt
def add_to_16(text):
if len(text.encode('utf-8')) % 16:
add = 16 - (len(text.encode('utf-8')) % 16)
else:
add = 0
text = text + ('\0' * add)
return text.encode('utf-8')
# 加密函数
def encrypt(text):
key = '9999999999999999'.encode('utf-8')
mode = AES.MODE_ECB
text = add_to_16(text)
cryptos = AES.new(key, mode)
cipher_text = cryptos.encrypt(text)
return b2a_hex(cipher_text)
# 解密后,去掉补足的空格用strip() 去掉
def decrypt(text):
key = '9999999999999999'.encode('utf-8')
mode = AES.MODE_ECB
cryptor = AES.new(key, mode)
plain_text = cryptor.decrypt(a2b_hex(text))
return bytes.decode(plain_text).rstrip('\0')
if __name__ == '__main__':
e = DES_encrypt("hello world") # 加密
print(type(e))
d = DES_decrypt(e) # 解密
print("加密:", e)
print("解密:", d) | 22.904762 | 51 | 0.626819 | [
"MIT"
] | peterzheng98/Valentine-Gift | try.py | 1,024 | Python |
"""This module contains the general information for ChassisPowerMonitor ManagedObject."""
from ...imcmo import ManagedObject
from ...imccoremeta import MoPropertyMeta, MoMeta
from ...imcmeta import VersionMeta
class ChassisPowerMonitorConsts:
pass
class ChassisPowerMonitor(ManagedObject):
"""This is ChassisPowerMonitor class."""
consts = ChassisPowerMonitorConsts()
naming_props = set([])
mo_meta = {
"modular": MoMeta("ChassisPowerMonitor", "chassisPowerMonitor", "pwrmonitor", VersionMeta.Version2013e, "OutputOnly", 0xf, [], ["admin", "read-only", "user"], ['equipmentChassis'], [], ["Get"])
}
prop_meta = {
"modular": {
"average": MoPropertyMeta("average", "average", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, None, None, None, None, [], []),
"child_action": MoPropertyMeta("child_action", "childAction", "string", VersionMeta.Version2013e, MoPropertyMeta.INTERNAL, None, None, None, None, [], []),
"current": MoPropertyMeta("current", "current", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, None, None, None, None, [], []),
"dn": MoPropertyMeta("dn", "dn", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, 0x2, 0, 255, None, [], []),
"maximum": MoPropertyMeta("maximum", "maximum", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, None, None, None, None, [], []),
"minimum": MoPropertyMeta("minimum", "minimum", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, None, None, None, None, [], []),
"period": MoPropertyMeta("period", "period", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, None, None, None, None, [], []),
"rn": MoPropertyMeta("rn", "rn", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, 0x4, 0, 255, None, [], []),
"status": MoPropertyMeta("status", "status", "string", VersionMeta.Version2013e, MoPropertyMeta.READ_ONLY, 0x8, None, None, r"""((removed|created|modified|deleted),){0,3}(removed|created|modified|deleted){0,1}""", [], []),
},
}
prop_map = {
"modular": {
"average": "average",
"childAction": "child_action",
"current": "current",
"dn": "dn",
"maximum": "maximum",
"minimum": "minimum",
"period": "period",
"rn": "rn",
"status": "status",
},
}
def __init__(self, parent_mo_or_dn, **kwargs):
self._dirty_mask = 0
self.average = None
self.child_action = None
self.current = None
self.maximum = None
self.minimum = None
self.period = None
self.status = None
ManagedObject.__init__(self, "ChassisPowerMonitor", parent_mo_or_dn, **kwargs)
| 42.835821 | 234 | 0.619164 | [
"Apache-2.0"
] | CiscoUcs/imcsdk | imcsdk/mometa/chassis/ChassisPowerMonitor.py | 2,870 | Python |
"""
FILE : BiLSTM.py
FUNCTION : None
"""
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.nn.utils.rnn import pack_padded_sequence, pad_packed_sequence
import random
from DataUtils.Common import *
from models.initialize import *
from models.modelHelp import prepare_pack_padded_sequence
torch.manual_seed(seed_num)
random.seed(seed_num)
class BiLSTM(nn.Module):
"""
BiLSTM
"""
def __init__(self, **kwargs):
super(BiLSTM, self).__init__()
for k in kwargs:
self.__setattr__(k, kwargs[k])
V = self.embed_num
D = self.embed_dim
C = self.label_num
paddingId = self.paddingId
self.embed = nn.Embedding(V, D, padding_idx=paddingId)
if self.pretrained_embed:
self.embed.weight.data.copy_(self.pretrained_weight)
else:
init_embedding(self.embed.weight)
self.dropout_embed = nn.Dropout(self.dropout_emb)
self.dropout = nn.Dropout(self.dropout)
self.bilstm = nn.LSTM(input_size=D, hidden_size=self.lstm_hiddens, num_layers=self.lstm_layers,
bidirectional=True, batch_first=True, bias=True)
self.linear = nn.Linear(in_features=self.lstm_hiddens * 2, out_features=C, bias=True)
init_linear(self.linear)
def forward(self, word, sentence_length):
"""
:param word:
:param sentence_length:
:param desorted_indices:
:return:
"""
word, sentence_length, desorted_indices = prepare_pack_padded_sequence(word, sentence_length, device=self.device)
x = self.embed(word) # (N,W,D)
x = self.dropout_embed(x)
packed_embed = pack_padded_sequence(x, sentence_length, batch_first=True)
x, _ = self.bilstm(packed_embed)
x, _ = pad_packed_sequence(x, batch_first=True)
x = x[desorted_indices]
x = self.dropout(x)
x = torch.tanh(x)
logit = self.linear(x)
return logit
class BiLSTM(nn.Module):
def __init__(self, vocab_size, emb_size, hidden_size, out_size):
""":
vocab_size:
emb_size:
hidden_size:
out_size:
"""
super(BiLSTM, self).__init__()
self.embedding = nn.Embedding(vocab_size, emb_size)
self.bilstm = nn.LSTM(emb_size, hidden_size,
batch_first=True,
bidirectional=True)
self.lin = nn.Linear(2*hidden_size, out_size)
def forward(self, sents_tensor, lengths):
emb = self.embedding(sents_tensor) # [B, L, emb_size]
packed = pack_padded_sequence(emb, lengths, batch_first=True)
rnn_out, _ = self.bilstm(packed)
# rnn_out:[B, L, hidden_size*2]
rnn_out, _ = pad_packed_sequence(rnn_out, batch_first=True)
scores = self.lin(rnn_out) # [B, L, out_size]
return scores
def test(self, sents_tensor, lengths, _):
logits = self.forward(sents_tensor, lengths) # [B, L, out_size]
_, batch_tagids = torch.max(logits, dim=2)
return batch_tagids
| 30.892157 | 121 | 0.623294 | [
"Apache-2.0"
] | Ahmed2xD/NER-with-bilstm-CRF-CNN | models/BiLSTM.py | 3,155 | Python |
'''
@author: kris
'''
# import modules; set up logging
from gensim.models import Word2Vec
from gensim.models import KeyedVectors
from gensim.test.utils import datapath
import numpy as np
import logging, os, sys, gzip
import datetime
logging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', filename='word2vec.out', level=logging.INFO)
# Path to a file that contains lines with the locations of files
# containing the sentences we want for our Word2Vec model
# Also works with entities that are just stacked line by line
pathsLocator = "./sentencesPaths.txt"
outputPath = "./entity_embeddings.txt"
# Model to load
to_load = '/vol2/cb/crunchbase-201806/embeddings/dim200-iter10-win5/CB_sg1_size200_mincount1_window5_neg15_iter10.wv.vectors.npy'
#'/home/faerberm/mag-training/MAG_sg1_size128_minCount5_window5_neg15_iter10_alpha_cbowMean.wv.vectors.npy'
#'/vol2/cb/crunchbase-201806/embeddings/dim200-iter10-win5/CB_sg1_size200_mincount1_window5_neg15_iter10'
#'MAG_sg1_size128_minCount5_window5_neg15_iter5'
loadKeyedVector = True
#'dbpedia_sg1_size200_mincount1_window5_neg15_iter10'
#'RDF2Vec_sg1_size200_mincount1_window5_neg15_iter20'
#'MAG_sg1_size200_mincount1_window5_neg15_iter15'
#What is the newline character on the machine
newline = '\n'
ignorePrefix = '#'
#What separates one walk from another (aka. one sentence from another)?
walkSeparator = "\t"
#What separates the single 'units' of a given walk?
hopSeparator = '->'
# Mapping dict
entity_mapping_dict = {}
# Mapping file
mapping_file = "/home/noulletk/prog/bmw/dbpedia_full/resources/data/walks/walk_entity_mapping.txt"
mapping_sep = "\t"
hasMapping = False
iterationCounter = {'val': 0}
#Load mappings if there are any
if hasMapping:
for mapping_line in open(mapping_file, mode='rt'):
mapping_tokens = mapping_line.rstrip(newline).split(mapping_sep)
if len(mapping_tokens) == 2:
entity_mapping_dict[mapping_tokens[0]] = mapping_tokens[1]
print("Loaded %s mappings!" % (len(entity_mapping_dict)))
class MySentences:
def __init__(self, iterationCounter):
self.iterationCounter = iterationCounter
def __iter__(self):
print("Running Iteration #%s" % (iterationCounter['val']))
iterationCounter['val'] += 1
# Iterate to find which files are to be read
for fname in open(pathsLocator, mode='rt'): # os.listdir(self.dirname):
sentencesPath = fname.rstrip(newline)
# Ignore commented-out lines
if sentencesPath.startswith(ignorePrefix):
continue
now = datetime.datetime.now()
print("[%s] Grabbing sentences from: %s" % (now.strftime("%Y-%m-%d %H:%M"), sentencesPath))
try:
# Go through all paths
for line in open(sentencesPath, mode='rt'):
# If you're NOT grouping the walks and separating them by tabs
sentence = line.rstrip(newline).split(hopSeparator)
for tokenPos in range(len(sentence)):
token = sentence[tokenPos]
# Give the proper URL for the entity IF it exists, otherwise return the entity itself
sentence[tokenPos] = entity_mapping_dict.get(token, token)
#print(sentence)
yield sentence
except Exception:
print("Failed reading file:")
print(sentencesPath)
#load model
if loadKeyedVector:
print("Loading [KeyedVectors] from: ",to_load)
#model_wv = KeyedVectors.load(to_load, mmap='r')
#model_wv = KeyedVectors.load_word2vec_format(to_load, binary=True)
#model_wv = KeyedVectors.load_word2vec_format(to_load)
model_wv = KeyedVectors.load(to_load)
#model_wv = KeyedVectors.load_word2vec_format(datapath('word2vec_pre_kv_c'), binary=False) # C text format
#model_wv = KeyedVectors.load_word2vec_format(to_load, binary=True, unicode_errors='ignore')
else:
print("Loading [MODEL] from: ",to_load)
model_wv = Word2Vec.load(to_load).wv
print("Vocab keys size:",len(model_wv.vocab.keys()))
print("Outputting entity embeddings to: ",outputPath)
sentences = MySentences(iterationCounter)
#Open the output file for the entity embeddings
outFile = open(outputPath, "w")
#Make a dictionary for in-memory aggregation while going over sentences
default_val = None
entity_embeddings_dict = {}
vocab_keys = model_wv.vocab.keys()
displayCounter = 0
maxDisplay = 10
for voc in vocab_keys:
print(voc)
if displayCounter >= maxDisplay:
break
displayCounter+=1
print("Compute entity embeddings (through combination of word embeddings)...")
counter = 0
'''
for sentence in sentences:
entity = sentence[0]
entity_embedding = None
#Sum over all words' embeddings and then output the resulting embedding
for word in sentence:
word_embedding = model.wv[word]
if default_val is None:
#Initialise default_val if it isn't yet
default_val = np.zeros(word_embedding.shape)
if entity_embedding is None:
entity_embedding = np.zeros(word_embedding.shape)
entity_embedding += word_embedding
entity_embeddings_dict[entity] = entity_embeddings_dict.get(entity, default_val) + entity_embedding
if (counter % 1000000 == 0):
print("Combined word embeddings: ",counter)
print("Last one completed: ",entity)
counter+=1
'''
#Go through all sentences to see which entities we want
for sentence in sentences:
# idea is that the entity is in the document, so we check what it is like and
# since every entity has 'the same' treatment, that we can determine their probabilities based on that
entity = sentence[0]
if hasMapping:
entity = entity_mapping_dict.get(entity, entity)
entity_embedding = None
dict_val = entity_embeddings_dict.get(entity, None)
if (dict_val is None):
if entity in vocab_keys:
entity_embedding = model_wv[entity]
entity_embeddings_dict[entity] = entity_embedding
#Encountered first time, so output it
outFile.write("%s" % entity)
for number in entity_embedding:
outFile.write("\t%s" % number)
outFile.write("\n")
if (counter % 1000000 == 0):
print("Lines passed through: ",counter)
print("Current line's entity: ",entity)
print("Embeddings output: ",len(entity_embeddings_dict))
counter+=1
#print("Output computed entity embeddings!")
#for (entity, entity_embedding) in entity_embeddings_dict.items():
# #Output computed embedding
# outFile.write("%s" % entity)
# for number in entity_embedding:
# outFile.write("\t%s" % number)
# outFile.write("\n")
#Close the output file post finishing output operations
outFile.close()
print("Finished outputting entity embeddings")
| 34.737705 | 129 | 0.757747 | [
"MIT"
] | michaelfaerber/Agnos | scripts/loadModelDoEntityEmbeddingsUnsorted.py | 6,357 | Python |
"""inter-base steganography
producing base32 and base64 decodable strings"""
from base64 import b64encode, b64decode
import string
from itertools import product
from argparse import ArgumentParser
CHARSET = string.printable.encode()
B32_CHARSET = (string.ascii_uppercase + '234567').encode()
B64_CHARSET = (
string.ascii_lowercase +
string.ascii_uppercase +
string.digits +
'+/').encode()
ASCII_LOWER = string.ascii_lowercase.encode()
WHITESPACE = string.whitespace.encode()
ALPHA_SPACE = (
string.ascii_uppercase +
string.ascii_lowercase +
string.whitespace).encode()
ASCII_SUBS = {"a": ["a", "A", "4", "@"],
"b": ["b", "B", "8", "6"],
"c": ["c", "C", "("],
"d": ["d", "D"],
"e": ["e", "E", "3"],
"f": ["f", "F"],
"g": ["g", "G", "6", "9"],
"h": ["h", "H", "#"],
"i": ["i", "I", "1", "|", "!"],
"j": ["j", "J", "]", ";"],
"k": ["k", "K"],
"l": ["l", "L", "1", "|"],
"m": ["m", "M"],
"n": ["n", "N"],
"o": ["o", "O", "0"],
"p": ["p", "P"],
"q": ["q", "Q", "9"],
"r": ["r", "R", "2"],
"s": ["s", "S", "5", "$"],
"t": ["t", "T", "7", "+"],
"u": ["u", "U"],
"v": ["v", "V"],
"w": ["w", "W"],
"x": ["x", "X"],
"y": ["y", "Y"],
"z": ["z", "Z", "2", "%"],
"0": ["0"],
"1": ["1"],
"2": ["2"],
"3": ["3"],
"4": ["4"],
"5": ["5"],
"6": ["6"],
"7": ["7"],
"8": ["8"],
"9": ["9"],
" ": [" ", "\t", "_"]
}
def all_variations(word: str) -> list:
"""
Produce all single-character leet variations of a string
"""
ans = [""]
for leet_letter in [ASCII_SUBS[i] for i in word]:
ans = [x + y for x in ans for y in leet_letter]
return ans
def variation_gen(word: str):
"""
Produces all single-character leet variations of a string
Args:
word: a 3 character string to generate all variations
Returns:
generator: generator for all possible leet variations
"""
return product(*(ASCII_SUBS[i] for i in word))
def all_valid_variations(word: str) -> list:
"""
Returns all leet variations of a triplet which result in a
Base32 only charset words on base64 encoding
Args:
word: An english triplet
Returns:
list: of all valid variations
"""
result = []
for variation in variation_gen(word):
if all(i in B32_CHARSET for i in b64encode(
''.join(variation).encode())):
result.append("".join(variation))
return result
def valid_variation(word: str) -> str:
"""
Generates a single valid variation
Args:
word: the triplet to generate a variation from
Returns:
str: A valid variation of `word` or None otherwise
"""
for variation in variation_gen(word):
if all(i in B32_CHARSET for i in b64encode(
''.join(variation).encode())):
return "".join(variation)
return None
# List to precompute the triplets for which there doesnt exist a valid
# variation
NON_LEET = []
for perm in product(string.ascii_lowercase + ' ' + string.digits, repeat=3):
if not valid_variation(''.join(perm)):
NON_LEET.append(''.join(perm))
def transform(strng: str) -> str:
"""
Transform the string to only lower alpha and numerics and spaces
Converts uppercase to lower case and strips all other characters except
space
"""
for char in string.punctuation + string.whitespace[1:]:
strng = strng.replace(char, '')
return strng.lower() + ' ' * (8 - len(strng) % 8)
def master_encode(strng: str) -> bytes:
"""
Encodes a string to its leet equivalent (sans punctuation) which when
base64 encoded contains only base32 characters
"""
if isinstance(strng, (bytes, bytearray)):
strng = strng.decode()
strng = transform(strng)
result = ''
i = 0
while i < len(strng):
try:
current = strng[i:i + 3]
if current in NON_LEET:
if current[:2] + ' ' not in NON_LEET:
result += valid_variation(current[:2] + ' ')
i += 2
elif current[0] + ' ' not in NON_LEET:
result += valid_variation(current[0] + ' ')
i += 1
elif ' {} '.format(current[0]) not in NON_LEET:
result += valid_variation(' {} '.format(current[0]))
i += 1
elif ' {}'.format(current[0]) not in NON_LEET:
result += valid_variation(' {}'.format(current[0]))
i += 1
else:
i += 1
else:
result += valid_variation(current)
i += 3
except TypeError:
i += 1
return b64encode(result.encode())
if __name__ == "__main__":
PARSER = ArgumentParser(description="")
PARSER.add_argument(
'--input',
help='read a single line directly from input',
action="store_true")
PARSER.add_argument(
'--show',
help='shows the transformed input which results in correct encoding',
action="store_true")
PARSER.add_argument(
'--file',
help='reading text from file for conversion',
action="append")
ARGS = PARSER.parse_args()
TEST_STRING = """Steganography is the practice of concealing a file,
message, image, or video within another file, message, image, or video.
The word steganography comes from Greek steganographia, which combines
the words steganos meaning "covered or concealed", and graphia meaning
"writing". The first recorded use of the term was by Johannes Trithemius
in his Steganographia, a treatise on cryptography and steganography,
disguised as a book on magic. Generally, the hidden messages appear to
be (or to be part of) something else: images, articles, shopping lists,
or some other cover text. For example, the hidden message may be in
invisible ink between the visible lines of a private letter. Some
implementations of steganography that lack a shared secret are forms
of security through obscurity, and key-dependent steganographic schemes
adhere to Kerckhoffs's principle."""
if ARGS.file:
with open(ARGS.file[0], 'rb') as inp_file:
TEST_STRING = inp_file.read()
else:
TEST_STRING = input("input the line to encode:\n")
ENCODED_STRING = master_encode(TEST_STRING)
print("ENCODED STRING: {}".format(ENCODED_STRING))
if ARGS.show:
print("Transformed string: {}".format(b64decode(ENCODED_STRING)))
# WTBVICAJV2VSZSBFWHBFY3RJIG4JOSBGTGFHNSBCVXQJYTFMICAJWTBVIDZFVCBJNSB3ZTFS\
# ZCBCYXNFNSBCYSAJTWJPMDJMZSAJTWVOVCBET25UICAJICB3T3JSWSBJVHMJIGYJVW4JIG4JZXZ\
# FIHIJVCNFTGVTNSAJ
| 33.747664 | 82 | 0.542924 | [
"MIT"
] | deut-erium/BASEic-steganography | encode.py | 7,222 | Python |
import sys
import datetime
def capitalize(string):
return string[0].upper() + string[1:]
action = sys.argv[1]
file_path = sys.argv[2]
project_name = sys.argv[3]
namespace = sys.argv[4]
now = datetime.datetime.now()
date = now.strftime("%m-%d-%Y %H:%M:%S")
args = sys.argv[6:]
username = "Logan Rickert"
def new_class():
file_name = sys.argv[5]
cpp_file_path = file_path + "src/" + file_name + ".cpp"
h_file_path = file_path + "include/" + file_name + ".h"
if len(args) % 2 != 0:
print "You must have an even amount of arguments!"
sys.exit()
parse = []
for arg in xrange(0,len(args),2):
parse.append([args[arg], args[arg + 1]])
cpp_file_contents = None
h_file_contents = None
with open(cpp_file_path, 'r') as f:
cpp_file_contents = f.read()
with open(h_file_path, 'r') as f:
h_file_contents = f.read()
cpp_file_contents = cpp_file_contents.replace(
"{{class_name}}", file_name
)
cpp_file_contents = cpp_file_contents.replace(
"{{namespace}}", namespace
)
cpp_file_contents = cpp_file_contents.replace(
"{{date}}", date
)
cpp_file_contents = cpp_file_contents.replace(
"{{username}}", username
)
if len(args) > 0:
construct_init = file_name + "::" + file_name + "("
for key, value in parse:
construct_init += key + " s" + capitalize(value) + ", "
construct_init = construct_init[:-2] + ") {"
cpp_file_contents = cpp_file_contents.replace(
"{{construct_init}}", construct_init
)
construct_init_equals = ""
for key, value in parse:
construct_init_equals += "\t" + value + " = s" + capitalize(value) + ";\n"
construct_init_equals += "}"
cpp_file_contents = cpp_file_contents.replace(
"{{construct_init_equals}}", construct_init_equals
)
getters_setters = ""
for key, value in parse:
getters_setters += """%s %s::get%s() {
return %s;
}
void %s::set%s(%s s%s) {
%s = s%s;
}
""" % (
key,
file_name,
capitalize(value),
value,
file_name,
capitalize(value),
key,
capitalize(value),
value,
capitalize(value)
)
getters_setters = getters_setters[:-2]
cpp_file_contents = cpp_file_contents.replace(
"{{getters_setters}}", getters_setters
)
else:
cpp_file_contents = cpp_file_contents.replace(
"\n{{construct_init}}\n", ""
)
cpp_file_contents = cpp_file_contents.replace(
"{{construct_init_equals}}\n", ""
)
cpp_file_contents = cpp_file_contents.replace(
"\n{{getters_setters}}\n", ""
)
with open(cpp_file_path, 'w') as f:
f.write(cpp_file_contents)
h_file_contents = h_file_contents.replace(
"{{class_name_caps}}", file_name.upper()
)
h_file_contents = h_file_contents.replace(
"{{class_name}}", file_name
)
h_file_contents = h_file_contents.replace(
"{{username}}", username
)
h_file_contents = h_file_contents.replace(
"{{namespace}}", namespace
)
h_file_contents = h_file_contents.replace(
"{{date}}", date
)
if len(args) > 0:
class_construct_full = file_name + "("
for key, value in parse:
class_construct_full += key + ", "
class_construct_full = class_construct_full[:-2] + ");"
h_file_contents = h_file_contents.replace(
"{{class_construct_full}}", class_construct_full
)
getters_setters = ""
for key, value in parse:
getters_setters += "\t\t" + key + " get" + capitalize(value) + "();\n"
getters_setters += '\n'
for key, value in parse:
getters_setters += "\t\tvoid set" + capitalize(value) + "(" + key + " s" + capitalize(value) + ");\n"
h_file_contents = h_file_contents.replace(
"{{getters_setters}}", getters_setters
)
class_fields = ""
for key, value in parse:
class_fields += "\t\t" + key + " " + value + ";\n"
h_file_contents = h_file_contents.replace(
"{{class_fields}}", class_fields
)
else:
h_file_contents = h_file_contents.replace(
"\n\t\t{{class_construct_full}}", ""
)
h_file_contents = h_file_contents.replace(
"{{getters_setters}}\n", ""
)
h_file_contents = h_file_contents.replace(
"{{class_fields}}", ""
)
with open(h_file_path, 'w') as f:
f.write(h_file_contents)
def new_main():
cpp_file_path = file_path + "/src/Main.cpp"
cpp_file_contents = None
h_file_contents = None
with open(cpp_file_path, 'r') as f:
cpp_file_contents = f.read()
cpp_file_contents = cpp_file_contents.replace(
"{{class_name}}", "Main"
)
cpp_file_contents = cpp_file_contents.replace(
"{{namespace}}", namespace
)
cpp_file_contents = cpp_file_contents.replace(
"{{username}}", username
)
cpp_file_contents = cpp_file_contents.replace(
"{{date}}", date
)
with open(cpp_file_path, 'w') as f:
f.write(cpp_file_contents)
if action == "class":
new_class()
elif action == "namespace" or action == "project":
new_main() | 20.828194 | 104 | 0.666244 | [
"CC0-1.0"
] | LoganRickert/CPP-Builder-And-Documentator | bin/parse_new_files.py | 4,728 | Python |
#!/usr/bin/python3
import time
import datetime
from gpiozero import InputDevice, LED
import subprocess
import requests
# RPI enumeration is:
# pin 5 & 6 are used for the button (3 & ground)
# pin 7 & 9 are used for the LED (4 & ground)
button_pin = 3
led_pin = 4
button = InputDevice(button_pin, pull_up=True)
last_active = False
last_press = None
led = LED(led_pin)
led.on()
def button_hold(now, seconds):
if seconds > 3:
print('button hold')
led.blink(.05, .5)
requests.get('http://localhost:8080/home')
time.sleep(2)
subprocess.call(['shutdown', '-h', 'now'], shell=False)
def button_release(now, seconds):
print('button release')
requests.get('http://localhost:8080/button')
while True:
cur_active = button.is_active
now = datetime.datetime.now()
if cur_active and not last_active:
last_press = now
if cur_active:
duration = now - last_press
button_hold(now, duration.total_seconds())
if not cur_active and last_active:
duration = now - last_press
button_release(now, duration.total_seconds())
last_active = cur_active
time.sleep(1/60) | 25.456522 | 63 | 0.668659 | [
"MIT"
] | kylemcdonald/bsp | pi/button/button.py | 1,171 | Python |
import argparse
import torch
from tqdm import tqdm
import vgg.data_loader.data_loaders as module_data
import vgg.model.loss as module_loss
import vgg.model.metric as module_metric
import vgg.model.model as module_arch
from vgg.parse_config import ConfigParser
def main(config):
logger = config.get_logger('test')
# setup data_loader instances
data_loader = getattr(module_data, config['data_loader']['type'])(
config['data_loader']['args']['data_dir'],
batch_size=512,
shuffle=False,
validation_split=0.0,
training=False,
num_workers=2
)
# build model architecture
model = config.init_obj('arch', module_arch)
logger.info(model)
# get function handles of loss and metrics
loss_fn = getattr(module_loss, config['loss'])
metric_fns = [getattr(module_metric, met) for met in config['metrics']]
logger.info('Loading checkpoint: {} ...'.format(config.resume))
checkpoint = torch.load(config.resume)
state_dict = checkpoint['state_dict']
if config['n_gpu'] > 1:
model = torch.nn.DataParallel(model)
model.load_state_dict(state_dict)
# prepare model for testing
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = model.to(device)
model.eval()
total_loss = 0.0
total_metrics = torch.zeros(len(metric_fns))
with torch.no_grad():
for i, (data, target) in enumerate(tqdm(data_loader)):
data, target = data.to(device), target.to(device)
output = model(data)
#
# save sample images, or do something with output here
#
# computing loss, metrics on test set
loss = loss_fn(output, target)
batch_size = data.shape[0]
total_loss += loss.item() * batch_size
for i, metric in enumerate(metric_fns):
total_metrics[i] += metric(output, target) * batch_size
n_samples = len(data_loader.sampler)
log = {'loss': total_loss / n_samples}
log.update({
met.__name__: total_metrics[i].item() / n_samples for i, met in enumerate(metric_fns)
})
logger.info(log)
if __name__ == '__main__':
args = argparse.ArgumentParser(description='PyTorch Template')
args.add_argument('-c', '--config', default=None, type=str,
help='config file path (default: None)')
args.add_argument('-r', '--resume', default=None, type=str,
help='path to latest checkpoint (default: None)')
args.add_argument('-d', '--device', default=None, type=str,
help='indices of GPUs to enable (default: all)')
config = ConfigParser.from_args(args)
main(config)
| 33.414634 | 93 | 0.641241 | [
"MIT"
] | mhd53/vgg-from-torch | vgg/test.py | 2,740 | Python |
import os
import argparse
from ops.os_operation import mkdir
import time
def write_slurm_sh_multi_H2(id,command_line, queue_name="learnfair",nodes=1,
gpu_per_node=8,wall_time=3*24*60,username="wang3702",CPU_PER_GPU=8):
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dependency_handler_path = os.path.join(os.getcwd(), "ops")
dependency_handler_path = os.path.join(dependency_handler_path, "handler.txt")
run_path = os.path.join(os.getcwd(), "log")
mkdir(run_path)
run_path = os.path.abspath(run_path)
prefix = "node%d_gpu%d"%(nodes,gpu_per_node)
batch_file = os.path.join(run_path, prefix+"slurm_job_" + str(id) + ".sh")
output_path = os.path.join(run_path, prefix+"output_" + str(id) + "_" + str(formatted_today + now) + ".log")
error_path = os.path.join(run_path, prefix+"error_" + str(id) + "_" + str(formatted_today + now) + ".log")
with open(batch_file, "w") as file:
file.write("#! /bin/bash\n")#!/bin/bash
file.write("#SBATCH --job-name=%s\n" % id)
file.write("#SBATCH --output=%s\n" % output_path)
file.write("#SBATCH --error=%s\n" % error_path)
file.write("#SBATCH --partition=%s\n"%queue_name)
file.write("#SBATCH --signal=USR1@600\n")
file.write("#SBATCH --nodes=%d\n" % nodes)
file.write("#SBATCH --ntasks-per-node=%d\n" % 1)
file.write("#SBATCH --mem=%dG\n"%(350/8*gpu_per_node))
file.write("#SBATCH --gpus=%d\n" % (nodes * gpu_per_node))
file.write("#SBATCH --gpus-per-node=%d\n" % (gpu_per_node))
file.write("#SBATCH --cpus-per-task=%d\n"%(CPU_PER_GPU*gpu_per_node))
file.write("#SBATCH --time=%d\n"%wall_time)
file.write("#SBATCH --mail-user=%[email protected]\n"%username)
file.write("#SBATCH --mail-type=FAIL\n")
file.write("#SBATCH --mail-type=end \n")
file.write('#SBATCH --constraint="volta"\n')
report_info = "%s job failed; \t" % id
report_info += "log path: %s; \t" % output_path
report_info += "error record path: %s\t" % error_path
report_info += "command line path: %s\t" % batch_file
file.write('#SBATCH --comment="%s"\n' % (report_info))
with open(dependency_handler_path, 'r') as rfile:
line = rfile.readline()
while line:
file.write(line)
line = rfile.readline()
file.write("export GLOO_SOCKET_IFNAME=\nexport NCCL_SOCKET_IFNAME=\n")
file.write("module load cuda/10.2 cudnn/v7.6.5.32-cuda.10.2 gcc/7.3.0\n")
#file.write("bash /private/home/wang3702/.bashrc\n")
#file.write("module load anaconda3\n")
file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
file.write("conda activate pytorch2\n")
file.write("master_node=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:4}\n")
file.write('dist_url="tcp://"\n')
file.write("dist_url+=$master_node\n")
file.write("dist_url+=:40000\n")
file.write("export MASTER_ADDR=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:4}\n")
file.write("export MASTER_PORT=29500\n")
file.write("srun --label "+command_line + " --slurm=1 --dist_url=$dist_url &\n")
file.write("wait $!\n")
file.write("set +x \n")
file.write("echo ..::Job Finished, but No, AGI is to BE Solved::.. \n")
# signal that job is finished
os.system('sbatch ' + batch_file)
def find_checkpoint(current_dir,checkpoint_name):
if not os.path.isdir(current_dir):
return None
listfiles = os.listdir(current_dir)
for item in listfiles:
sub_dir = os.path.join(current_dir,item)
if item==checkpoint_name:
return sub_dir
elif os.path.isdir(sub_dir):
search_result = find_checkpoint(sub_dir,checkpoint_name)
if search_result is not None:
return search_result
return None
def write_slurm_sh_multi(id,command_line, queue_name="learnfair",nodes=1,
gpu_per_node=8,wall_time=3*24*60,username="wang3702",
CPU_PER_GPU=8,gpu_memory=False,environment=0):
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dependency_handler_path = os.path.join(os.getcwd(), "ops")
dependency_handler_path = os.path.join(dependency_handler_path, "handler.txt")
run_path = os.path.join(os.getcwd(), "log")
mkdir(run_path)
run_path = os.path.abspath(run_path)
prefix = "node%d_gpu%d"%(nodes,gpu_per_node)
batch_file = os.path.join(run_path, prefix+"slurm_job_" + str(id) + ".sh")
output_path = os.path.join(run_path, prefix+"output_" + str(id) + "_" + str(formatted_today + now) + ".log")
error_path = os.path.join(run_path, prefix+"error_" + str(id) + "_" + str(formatted_today + now) + ".log")
with open(batch_file, "w") as file:
file.write("#! /bin/bash\n")#!/bin/bash
file.write("#SBATCH --job-name=%s\n" % id)
file.write("#SBATCH --output=%s\n" % output_path)
file.write("#SBATCH --error=%s\n" % error_path)
file.write("#SBATCH --partition=%s\n"%queue_name)
file.write("#SBATCH --signal=USR1@600\n")
file.write("#SBATCH --nodes=%d\n" % nodes)
file.write("#SBATCH --ntasks-per-node=%d\n" % 1)
file.write("#SBATCH --mem=%dG\n"%(350/8*gpu_per_node))#--mem : Specify the real memory required per node.
file.write("#SBATCH --gpus=%d\n" % (nodes * gpu_per_node))
file.write("#SBATCH --gpus-per-node=%d\n" % (gpu_per_node))
file.write("#SBATCH --cpus-per-task=%d\n"%(CPU_PER_GPU*gpu_per_node))
file.write("#SBATCH --time=%d\n"%wall_time)
file.write("#SBATCH --mail-user=%[email protected]\n"%username)
file.write("#SBATCH --mail-type=FAIL\n")
file.write("#SBATCH --mail-type=end \n")
if gpu_memory is False:
file.write('#SBATCH --constraint="volta"\n')
else:
file.write('#SBATCH --constraint="volta32gb"\n')
#file.write('#SBATCH --constraint="volta"\n')
report_info = "%s job failed; \t" % id
report_info += "log path: %s; \t" % output_path
report_info += "error record path: %s\t" % error_path
report_info += "command line path: %s\t" % batch_file
file.write('#SBATCH --comment="%s"\n' % (report_info))
with open(dependency_handler_path, 'r') as rfile:
line = rfile.readline()
while line:
file.write(line)
line = rfile.readline()
file.write("export GLOO_SOCKET_IFNAME=\nexport NCCL_SOCKET_IFNAME=\n")
file.write("module load cuda/10.2 cudnn/v7.6.5.32-cuda.10.2 gcc/7.3.0\n")
#file.write("bash /private/home/wang3702/.bashrc\n")
file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
#file.write("module load anaconda3\n")
file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
if environment==0:
file.write("conda activate pytorch2\n")
else:
file.write("conda activate pytorch\n")
file.write("master_node=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:4}\n")
file.write('dist_url="tcp://"\n')
file.write("dist_url+=$master_node\n")
file.write("dist_url+=:40000\n")
file.write("export MASTER_ADDR=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:4}\n")
file.write("export MASTER_PORT=29500\n")
file.write("srun --label "+command_line + " --slurm=1 --dist_url=$dist_url &\n")
file.write("wait $!\n")
file.write("set +x \n")
file.write("echo ..::Job Finished, but No, AGI is to BE Solved::.. \n")
# signal that job is finished
os.system('sbatch ' + batch_file)
def write_slurm_sh_multi2(id,command_line, queue_name="learnfair",nodes=1,
gpu_per_node=8,wall_time=3*24*60,username="wang3702",CPU_PER_GPU=8,
gpu_memory=False,environment=0):
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dependency_handler_path = os.path.join(os.getcwd(), "ops")
dependency_handler_path = os.path.join(dependency_handler_path, "handler.txt")
run_path = os.path.join(os.getcwd(), "log")
mkdir(run_path)
run_path = os.path.abspath(run_path)
prefix = "node%d_gpu%d"%(nodes,gpu_per_node)
batch_file = os.path.join(run_path, prefix+"slurm_job_" + str(id) + ".sh")
output_path = os.path.join(run_path, prefix+"output_" + str(id) + "_" + str(formatted_today + now) + ".log")
error_path = os.path.join(run_path, prefix+"error_" + str(id) + "_" + str(formatted_today + now) + ".log")
with open(batch_file, "w") as file:
file.write("#! /bin/bash\n")#!/bin/bash
file.write("#SBATCH --job-name=%s\n" % id)
file.write("#SBATCH --output=%s\n" % output_path)
file.write("#SBATCH --error=%s\n" % error_path)
file.write("#SBATCH --partition=%s\n"%queue_name)
file.write("#SBATCH --signal=USR1@600\n")
file.write("#SBATCH --nodes=%d\n" % nodes)
file.write("#SBATCH --ntasks-per-node=%d\n" % 1)
file.write("#SBATCH --mem=%dG\n"%(350/8*gpu_per_node))
file.write("#SBATCH --gpus=%d\n" % (nodes * gpu_per_node))
file.write("#SBATCH --gpus-per-node=%d\n" % (gpu_per_node))
file.write("#SBATCH --cpus-per-task=%d\n"%(CPU_PER_GPU*gpu_per_node))
file.write("#SBATCH --time=%d\n"%wall_time)
file.write("#SBATCH --mail-user=%[email protected]\n"%username)
file.write("#SBATCH --mail-type=FAIL\n")
file.write("#SBATCH --mail-type=end \n")
if gpu_memory is False:
file.write('#SBATCH --constraint="volta"\n')
else:
file.write('#SBATCH --constraint="volta32gb"\n')
report_info = "%s job failed; \t" % id
report_info += "log path: %s; \t" % output_path
report_info += "error record path: %s\t" % error_path
report_info += "command line path: %s\t" % batch_file
file.write('#SBATCH --comment="%s"\n' % (report_info))
with open(dependency_handler_path, 'r') as rfile:
line = rfile.readline()
while line:
file.write(line)
line = rfile.readline()
file.write("export GLOO_SOCKET_IFNAME=\nexport NCCL_SOCKET_IFNAME=\n")
file.write("module load cuda/10.2 cudnn/v7.6.5.32-cuda.10.2 gcc/7.3.0\n")
#file.write("bash /private/home/wang3702/.bashrc\n")
# file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
#file.write("module load anaconda3\n")
file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
if environment==0:
file.write("conda activate pytorch2\n")
else:
file.write("conda activate pytorch\n")
#file.write("source activate\n")
file.write("master_node=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:3}\n")
file.write('dist_url="tcp://"\n')
file.write("dist_url+=$master_node\n")
file.write("dist_url+=:40000\n")
file.write("export MASTER_ADDR=${SLURM_NODELIST:0:9}${SLURM_NODELIST:10:3}\n")
file.write("export MASTER_PORT=29500\n")
file.write("srun --label "+command_line + " &\n")
file.write("wait $!\n")
file.write("set +x \n")
file.write("echo ..::Job Finished, but No, AGI is to BE Solved::.. \n")
# signal that job is finished
os.system('sbatch ' + batch_file)
def write_slurm_sh_faster(id,command_line, queue_name="learnfair",nodes=1,
gpu_per_node=8,wall_time=3*24*60,username="wang3702",CPU_PER_GPU=8,
gpu_memory=False,environment=0):
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dependency_handler_path = os.path.join(os.getcwd(), "ops")
dependency_handler_path = os.path.join(dependency_handler_path, "handler.txt")
run_path = os.path.join(os.getcwd(), "log")
mkdir(run_path)
run_path = os.path.abspath(run_path)
batch_file = os.path.join(run_path, "slurm_job_" + str(id) + ".sh")
output_path = os.path.join(run_path, "output_" + str(id) + "_" + str(formatted_today + now) + ".log")
error_path = os.path.join(run_path, "error_" + str(id) + "_" + str(formatted_today + now) + ".log")
with open(batch_file, "w") as file:
file.write("#!/bin/bash\n")#!/bin/bash
file.write("#SBATCH --job-name=%s\n" % id)
file.write("#SBATCH --output=%s\n" % output_path)
file.write("#SBATCH --error=%s\n" % error_path)
file.write("#SBATCH --partition=%s\n"%queue_name)
file.write("#SBATCH --signal=USR1@600\n")
file.write("#SBATCH --nodes=%d\n" % nodes)
file.write("#SBATCH --ntasks-per-node=%d\n" % gpu_per_node)
file.write("#SBATCH --mem=%dG\n"%(int(350/8*gpu_per_node)))
file.write("#SBATCH --gpus=%d\n" % (nodes * gpu_per_node))
file.write("#SBATCH --cpus-per-task=%d\n"%(CPU_PER_GPU))
file.write("#SBATCH --time=%d\n"%wall_time)
file.write("#SBATCH --mail-user=%[email protected]\n"%username)
file.write("#SBATCH --mail-type=FAIL\n")
file.write("#SBATCH --mail-type=end \n")
if gpu_memory:
file.write('#SBATCH --constraint="volta32gb"\n')
else:
file.write('#SBATCH --constraint="volta"\n')
report_info = "%s job failed; \t" % id
report_info += "log path: %s; \t" % output_path
report_info += "error record path: %s\t" % error_path
report_info += "command line path: %s\t" % batch_file
file.write('#SBATCH --comment="%s"\n' % (report_info))
with open(dependency_handler_path, 'r') as rfile:
line = rfile.readline()
while line:
file.write(line)
line = rfile.readline()
file.write("module load cuda/10.2 cudnn/v7.6.5.32-cuda.10.2 gcc/7.3.0\n")
#file.write("bash /private/home/wang3702/.bashrc\n")
file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
#file.write("module load anaconda3\n")
file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
if environment==0:
file.write("conda activate pytorch2\n")
else:
file.write("conda activate pytorch\n")
#file.write("source activate\n")
file.write(command_line + " &\n")
file.write("wait $!\n")
file.write("set +x \n")
file.write("echo ..::Job Finished, but No, AGI is to BE Solved::.. \n")
# signal that job is finished
os.system('sbatch ' + batch_file)
def write_slurm_sh(id,command_line, queue_name="learnfair",nodes=1,
gpu_per_node=8,wall_time=3*24*60,username="wang3702",CPU_PER_GPU=10):
"""
Args:
id: running id
command_line: command line
outlog_path: saving path
Returns:
"""
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dependency_handler_path = os.path.join(os.getcwd(),"ops")
dependency_handler_path = os.path.join(dependency_handler_path,"handler.txt")
run_path = os.path.join(os.getcwd(),"log")
mkdir(run_path)
run_path = os.path.abspath(run_path)
batch_file = os.path.join(run_path,"slurm_job_"+str(id)+".sh")
output_path = os.path.join(run_path,"output_"+str(id)+"_"+str(formatted_today+now)+".log")
error_path = os.path.join(run_path,"error_"+str(id)+"_"+str(formatted_today+now)+".log")
with open(batch_file,"w") as file:
file.write("#!/bin/sh\n")
file.write("#SBATCH --job-name=%s\n"%id)
file.write("#SBATCH --output=%s\n"%output_path)
file.write("#SBATCH --error=%s\n"%error_path)
file.write("#SBATCH --partition=%s\n"%queue_name)
file.write("#SBATCH --signal=USR1@600\n")
file.write("#SBATCH --nodes=%d\n"%nodes )
file.write("#SBATCH --ntasks-per-node=1\n")
file.write("#SBATCH --mem=350G\n")
file.write("#SBATCH --gpus=%d\n"%(nodes*gpu_per_node))
file.write("#SBATCH --gpus-per-node=%d\n" % (gpu_per_node))
file.write("#SBATCH --cpus-per-task=%d\n"%(CPU_PER_GPU*gpu_per_node))
file.write("#SBATCH --time=%d\n"%wall_time)
file.write("#SBATCH --mail-user=%[email protected]\n"%username)
file.write("#SBATCH --mail-type=FAIL\n")
file.write("#SBATCH --mail-type=end \n")
file.write('#SBATCH --constraint="volta"\n')
report_info ="%s job failed; \t"%id
report_info += "log path: %s; \t"%output_path
report_info += "error record path: %s\t"%error_path
report_info += "command line path: %s\t"%batch_file
file.write('#SBATCH --comment="%s"\n'%(report_info))
with open(dependency_handler_path,'r') as rfile:
line = rfile.readline()
while line:
file.write(line)
line = rfile.readline()
#file.write("bash /private/home/wang3702/.bashrc\n")
# file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
#file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
#file.write("module load anaconda3\n")
#file.write("conda activate pytorch2\n")
file.write("module load cuda/10.2 cudnn/v7.6.5.32-cuda.10.2 gcc/7.3.0\n")
file.write("/private/home/wang3702/anaconda3/bin/conda init\n")
file.write("CONDA_BASE=$(conda info --base) ; source $CONDA_BASE/etc/profile.d/conda.sh\n")
file.write("conda activate pytorch2\n")
file.write(command_line+" &\n")
file.write("wait $!\n")
file.write("set +x \n")
file.write("echo ..::Job Finished, but No, AGI is to BE Solved::.. \n")
# signal that job is finished
os.system('sbatch ' + batch_file)
parser = argparse.ArgumentParser(description='slurm job submission')
parser.add_argument('--data', default="imagenet", type=str, metavar='DIR',
help='path to dataset')
parser.add_argument("--mode",type=int,default=0,help="control mode for training")
parser.add_argument("--type",type=int,default=0,help="running type control")
parser.add_argument("--roi",type=int,default = 20, help="number of rois sampled here")
parser.add_argument("--queue",type=int,default=0, help="queue specified list")
parser.add_argument("-F",type=str, default=None, help="resume path for running again")
parser.add_argument("--comment", type=str,default=None,help="adding comment for script names")
parser.add_argument("--node",type=int,default=1,help="nodes needed for training")
parser.add_argument("--gpu",type=int,default=8,help="number of gpus per node")
args = parser.parse_args()
if args.queue ==0:
queue_name = "learnfair"
elif args.queue ==1:
queue_name = "dev"
elif args.queue ==2:
queue_name = "scavenge"
elif args.queue ==3:
queue_name = 'priority'
elif args.queue ==4:
queue_name = 'learnlab'
elif args.queue==5:
queue_name = 'devlab'
elif args.queue==6:
queue_name = 'prioritylab'
dump_path= os.path.join(os.getcwd(),"swav_dump_100")
from ops.os_operation import mkdir
mkdir(dump_path)
import time
import datetime
today = datetime.date.today()
formatted_today = today.strftime('%y%m%d')
now = time.strftime("%H:%M:%S")
dump_path = os.path.join(dump_path, formatted_today + now)
if args.mode==1:
if args.type==0:
# command_line = "python3 main_adco.py --mode=1 --lr=0.06 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0006 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=57" % args.data
# write_slurm_sh("baseline_sym_moco_lr0.06_proj", command_line, queue_name)
command_line = "python3 main_adco.py --mode=1 --lr=0.06 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0006 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=16 --mask_size=32 " \
"--num_roi=1 " % args.data
write_slurm_sh("baseline_sym_moco_lr0.06", command_line, queue_name)
# command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=16 --mask_size=32 " \
# "--num_roi=1 --img_size=96 " % args.data
# write_slurm_sh("baseline_sym_moco_input96", command_line, queue_name)
#running all the baseline with 100 epochs
#base line moco
# command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=35 --mask_size=32 " \
# " --num_roi=1 " % args.data
# write_slurm_sh("baseline_sym_mocobn_100", command_line, queue_name)
# #moco multi baseline
# command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=18 --nmb_crops 2 6 " \
# "--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " % (args.data)
# write_slurm_sh("multi_moco_baseline_100_new", command_line, queue_name)
# # #moco multi sym baseline
# command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=20 --nmb_crops 4 " \
# "--size_crops 224 --min_scale_crops 0.14 --max_scale_crops 1.0 " % (args.data)
# write_slurm_sh("2key_multi_moco_baseline_4_224", command_line, queue_name)
# #swav multi baseline
# command_line = "python3 main_adco.py --mode=5 --type=0 --data=%s --epochs 100 --lr=0.6 " \
# "--lr_final 0.0006 --batch_size=256 --warmup_epochs 0 --freeze_prototypes_niters 5005 " \
# "--queue_length 3840 --epoch_queue_starts 15 --dist_url=tcp://localhost:10031 " \
# "--knn_batch_size=256 --cos=1 --momentum=0.9 --weight_decay=1e-6 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 --moco_k=3000 --moco_t=0.1 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --nmb_crops 2 " \
# "--size_crops 224 --min_scale_crops 0.14 --max_scale_crops 1.0 --dump_path %s " % (args.data,dump_path)
# write_slurm_sh("swav_baseline_100_only224", command_line, queue_name)
# command_line = "python3 main_adco.py --mode=5 --type=0 --data=%s --epochs 100 --lr=0.6 " \
# "--lr_final 0.0006 --batch_size=256 --warmup_epochs 0 --freeze_prototypes_niters 5005 " \
# "--queue_length 3840 --epoch_queue_starts 15 --dist_url=tcp://localhost:10031 " \
# "--knn_batch_size=256 --cos=1 --momentum=0.9 --weight_decay=1e-6 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 --moco_k=3000 --moco_t=0.1 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --nmb_crops 2 6 " \
# "--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 --dump_path %s " % (
# args.data, dump_path)
# write_slurm_sh("swav_baseline_100", command_line, queue_name)
elif args.type==10:
#half dropout results
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=10 " % args.data
if args.F is not None:
resume_name = os.path.split(os.path.abspath(args.F))[1]
command_line += "--resume=%s"%args.F
write_slurm_sh("halfdropoutnew_resume%s"%resume_name, command_line, queue_name)
else:
write_slurm_sh("halfdropoutnew", command_line, queue_name)
elif args.type==11:
# to make sure overlap region can really not work
for mask_size in [96, 160]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=11 --shift_ratio=0 " \
" --mask_size=%d " % (args.data,mask_size)
write_slurm_sh("type11_roimatch_%s"%mask_size, command_line, queue_name)
elif args.type==13:
for mask_size in [96,160]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=13 " \
"--mask_size=%d "%(args.data,mask_size)
write_slurm_sh("type13_singleroi_vs_global_%d"%mask_size,command_line,queue_name)
time.sleep(1)
elif args.type==14:
#roi vs global
for mask_size in [96,160]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=14 " \
"--mask_size=%d "%(args.data,mask_size)
write_slurm_sh("type14_singleroi_vs_global_%d"%mask_size,command_line,queue_name)
elif args.type==16:
for mask_size in [96,128,160]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=16 " \
"--mask_size=%d --num_roi=10 "%(args.data,mask_size)
write_slurm_sh("type16_roi+global_vs_global_%d"%mask_size,command_line,queue_name)
elif args.type==-16:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=16 --mask_size=32 --num_roi=1 " % args.data
if args.F is not None:
resume_name = os.path.split(os.path.abspath(args.F))[1]
command_line += " --resume=%s"%args.F
write_slurm_sh("baseline_sym_moco_resume%s"%resume_name, command_line, queue_name)
else:
write_slurm_sh("baseline_sym_moco", command_line,queue_name)
elif args.type==17:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=17 --mask_size=32" \
" --num_roi=%d" % (args.data,args.roi)
write_slurm_sh("type17_randroi_%d"%args.roi, command_line,queue_name)
elif args.type==-17:
#roi vs roi,with global as negative
for roi in [10,20,50,100]:
for mask_size in [32, 96, 160, 196]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=17 --mask_size=%d" \
" --num_roi=%d" % (args.data,mask_size, roi)
write_slurm_sh("type17_randroi_%d_masksize_%d" % (roi,mask_size), command_line,queue_name)
elif args.type==18:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=18 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 "% (args.data)
if args.F is not None:
resume_name = os.path.split(os.path.abspath(args.F))[1]
command_line += "--resume=%s"%args.F
write_slurm_sh("multi_moco_baseline_resume%s"%resume_name, command_line, queue_name)
else:
write_slurm_sh("multi_moco_baseline" , command_line, queue_name)
elif args.type==19:
for roi in [20]:
for mask_size in [32,160]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=19 --mask_size=%d" \
" --num_roi=%d" % (args.data,mask_size, roi)
write_slurm_sh("type19_randroi_%d_masksize_%d" % (roi,mask_size), command_line,queue_name)
elif args.type==20:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=20 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 "% (args.data)
if args.F is not None:
resume_name = os.path.split(os.path.abspath(args.F))[1]
command_line += " --resume=%s"%args.F
write_slurm_sh("2key_multi_moco_baseline_correct_resume%s"%resume_name, command_line, queue_name)
else:
write_slurm_sh("2key_multi_moco_baseline_correct", command_line, queue_name)
elif args.type==21:
for roi in [20]:
for mask_size in [96]:
command_line = "python3 main_adco.py --mode=1 --lr=0.09 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=768 --knn_batch_size=256 --cos=1 --lr_final=0.0009 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=21 --mask_size=%d" \
" --num_roi=%d" % (args.data,mask_size, roi)
write_slurm_sh("type21_randroi_%d_masksize_%d" % (roi,mask_size), command_line,queue_name)
elif args.type==22:
for roi in [50]:
for mask_size in [96]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=22 --mask_size=%d" \
" --num_roi=%d" % (args.data, mask_size, roi)
write_slurm_sh("type22_randroi_%d_masksize_%d" % (roi,mask_size), command_line,queue_name)
elif args.type==23:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=23 --nmb_crops 2 2 2 2 2 2 2 2" \
" --size_crops 96 112 128 144 160 176 192 208 " % args.data
write_slurm_sh("type23_specifyroi", command_line, queue_name)
elif args.type==-23:
# command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=200 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=23 --nmb_crops 6" \
# " --size_crops 96 " % args.data
# write_slurm_sh("type23_specifyroi_6_96", command_line, queue_name)
min_scale = 64
max_scale = 224
divide_list = [2,4,8,16,32]
pick_times = [1,2,3]
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale<max_scale:
check_list+=str(current_scale)+" "
num_list+=str(pick_time)+" "
current_scale+=divide
print(check_list)
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=200 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=23 --nmb_crops %s " \
" --size_crops %s " % (args.data,num_list,check_list)
write_slurm_sh("type23_specifyroi_%d_%d"%(pick_time,divide), command_line, queue_name)
elif args.type==24:
for alpha in [0.5, 1.0, 2.0]:
for local_t in [0.1,0.2,0.3]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=24 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=1.0 " % (args.data,local_t)
write_slurm_sh("type24_lg_t_%.3f_alpha_%.2f"%(local_t,alpha), command_line, queue_name)
elif args.type==25:
for alpha in [0.5]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=24 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % (args.data, local_t,alpha)
write_slurm_sh("type25_lgq_t_%.3f_alpha_%.2f" %(local_t,alpha), command_line, queue_name)
elif args.type==26:
for alpha in [0.5,1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=26 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % (args.data, local_t,alpha)
write_slurm_sh("type26_lgq_t_%.3f_alpha_%.2f" %(local_t,alpha), command_line, queue_name)
elif args.type == 27:
min_scale = 96
max_scale = 224
divide_list = [16]
pick_times = [1]
for learning_rate in [0.05]:#[0.02,0.03,0.04,0.05,0.06,0.1,0.15]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0.1,0.15,0.2,0.3]:#[0.3, 0.5, 1.0]:
for local_t in [0.12,0.15,0.18]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=27 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f " % \
(learning_rate,args.data, local_t,num_list, check_list, local_t, alpha)
write_slurm_sh("type27_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, pick_time, divide,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == -270:
for num_roi in [6,10,20,30]:
for crop_size in [64, 96, 128, 160, 192]:
for learning_rate in [0.05]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.18]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=27 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh(
"type27crop_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==-271:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.18,0.2]:
for moco_dim in [256,512]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=%d " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=27 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data,moco_dim, local_t, num_roi, crop_size, local_t, alpha)
write_slurm_sh(
"type27dim_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f_dim%d" % (
local_t, alpha, num_roi, crop_size, learning_rate,moco_dim),
command_line, queue_name)
time.sleep(1)
elif args.type == -27:
#calculate baseline 6*96 for type 27 as a direct cmp with SWAV
for learning_rate in [0.05]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.18]:
for moco_dim in [128,256,512]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=27 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type27baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 28:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=28 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " % (args.data)
write_slurm_sh("type28_small_inside", command_line, queue_name)
elif args.type==29:
for learning_rate in [0.03]:
for alpha in [0.5,1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.2f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.5f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=29 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " \
"" % (learning_rate,args.data, learning_rate/100,local_t, alpha)
write_slurm_sh("type29_lgq_t_%.3f_alpha_%.2f_lr_%.4f" % (local_t, alpha,learning_rate), command_line, queue_name)
elif args.type==30:
for learning_rate in [0.03]:
for alpha in [0.5,1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.2f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.5f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=30 --nmb_crops 6 " \
" --size_crops 96 --local_t=%.4f --alpha=%.2f " \
"" % (learning_rate,args.data, learning_rate/100,local_t, alpha)
write_slurm_sh("type30_lgq_t_%.3f_alpha_%.2f_lr_%.4f" % (local_t, alpha,learning_rate), command_line, queue_name)
elif args.type==31:
for learning_rate in [0.03]:
for alpha in [0.5]:
for local_t in [0.2]:
for num_roi in [5, 10, 20]:
for mask_size in [96]:
command_line = "python3 main_adco.py --mode=1 --lr=%.2f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.5f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=31 " \
"--local_t=%.4f --alpha=%.2f --num_roi=%d --mask_size=%d " \
"" % (learning_rate, args.data, learning_rate / 100,
local_t, alpha,num_roi,mask_size)
write_slurm_sh("type31_lgq_t_%.3f_alpha_%.2f_lr_%.4f_roi%d_mask%d" %
(local_t, alpha, learning_rate,num_roi,mask_size),
command_line, queue_name)
elif args.type==32:
for learning_rate in [0.03]:
for alpha in [0.5]:
for local_t in [0.2]:
for num_roi in [5, 10, 20]:
for mask_size in [96]:
command_line = "python3 main_adco.py --mode=1 --lr=%.2f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.5f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=32 " \
"--local_t=%.4f --alpha=%.2f --num_roi=%d --mask_size=%d " \
"" % (learning_rate, args.data, learning_rate / 100,
local_t, alpha,num_roi,mask_size)
write_slurm_sh("type32_lgq_t_%.3f_alpha_%.2f_lr_%.4f_roi%d_mask%d" %
(local_t, alpha, learning_rate,num_roi,mask_size),
command_line, queue_name)
elif args.type==33:
for learning_rate in [0.03,0.04,0.05,0.06,0.09,0.12]:
for alpha in [0.5,1.0,2.0,5.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=33 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh("multimoco_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name)
elif args.type==-28:
for learning_rate in [0.06]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=28 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh("multimocoinside_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name)
elif args.type==34:
min_scale = 96
max_scale = 224
divide_list = [16]
pick_times = [1]
for learning_rate in [0.04, 0.05]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0.1, 0.3, 0.5,1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=34 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, num_list, check_list, local_t, alpha)
write_slurm_sh("type34_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, pick_time, divide, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 36:
min_scale = 96
max_scale = 224
divide_list = [16]
pick_times = [1]
for learning_rate in [0.04,0.05]:#[0.02,0.03,0.04,0.05,0.06,0.1,0.15]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0.1]:#[0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=36 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f " % \
(learning_rate,args.data, local_t,num_list, check_list, local_t, alpha)
write_slurm_sh("type36_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, pick_time, divide,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==37:
for learning_rate in [0.03,0.04,0.05,0.06]:
for alpha in [0.1,0.3,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=37 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type37baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==38:
min_scale = 96
max_scale = 224
divide_list = [16]
pick_times = [1]
for learning_rate in [0.05]: # [0.02,0.03,0.04,0.05,0.06,0.1,0.15]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0]: #[0.1, 0.3, 0.5, 1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=38 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,"", "", local_t, alpha)
write_slurm_sh("type38_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, pick_time, divide, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==-38:
for learning_rate in [0.05]:
for alpha in [0.1,0.3,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=38 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type38baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==39:
for learning_rate in [0.05]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=39 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type39baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==40:
for learning_rate in [0.05]:
for alpha in [0.5]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=40 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type40baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==41:
for mask_size in [96]:
command_line = "python3 main_adco.py --mode=1 --lr=0.03 --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=41 " \
"--mask_size=%d "%(args.data,mask_size)
write_slurm_sh("type41_singleroi_vs_global_%d"%mask_size,command_line,queue_name)
elif args.type==42:
for learning_rate in [0.05]:
for alpha in [0.1,0.5]: # [0.3, 0.5, 1.0]:
for local_t in [0.15,0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=42 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type42baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==43:
for learning_rate in [0.05]:
for alpha in [0.1,0.5]: # [0.3, 0.5, 1.0]:
for local_t in [0.15,0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=43 --nmb_crops 1 6" \
" --size_crops 224 96 --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t,local_t, alpha)
write_slurm_sh("type43baseline_lgq_t_%.3f_alpha_%.2f_6_96_lr%.4f" % (local_t, alpha,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 44:
# for num_roi in [6]:
# for crop_size in [96]:
# for learning_rate in [0.05]:
# for alpha in [0.1]: # [0.3, 0.5, 1.0]:
# for local_t in [0.15, 0.18, 0.2]:
# for sample_ratio in [2,4]:
# command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
# "--dist_url=tcp://localhost:10031 --epochs=100 " \
# "--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
# "--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
# "--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
# "--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
# "--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=44 --nmb_crops 1 %d" \
# " --size_crops 224 %d --local_t=%.4f --alpha=%.2f --sample_ratio=%d " % \
# (learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha,sample_ratio)
# write_slurm_sh(
# "type44crop_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f_ratio%d" % (local_t, alpha, num_roi,crop_size, learning_rate,sample_ratio),
# command_line, queue_name)
# time.sleep(1)
for num_roi in [6]:
for crop_size in [96,192]:
for learning_rate in [0.03,0.05,0.06]:
for alpha in [0.1,0.3,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=44 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh(
"type44_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==-44:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0.1,0.5]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=44 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh(
"type44align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==45 or args.type==46:
for crop_size in [96]:
for learning_rate in [0.03,0.04,0.05]:
for alpha in [0.1,0.3,0.5,1,2]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --mask_size %d" \
" --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t, args.type, crop_size,local_t, alpha)
write_slurm_sh(
"type%d_crop_lgq_t_%.3f_alpha_%.2f_%d_lr%.4f" % (args.type, local_t,alpha,
crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type ==47:
min_scale = 96
max_scale = 224
divide_list = [16]
pick_times = [1]
for learning_rate in [0.03,0.05]: # [0.02,0.03,0.04,0.05,0.06,0.1,0.15]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0.1,0.5,1.0]: # [0.1, 0.3, 0.5, 1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=47 " \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f " % \
(learning_rate, args.data, local_t, check_list, local_t, alpha)
write_slurm_sh("type47_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, pick_time, divide, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type ==49:
min_scale = 96
max_scale = 224
divide_list = [2,4,8,16,32]
pick_times = [1]
for learning_rate in [0.06]: # [0.02,0.03,0.04,0.05,0.06,0.1,0.15]:
for pick_time in pick_times:
for divide in divide_list:
check_list = ""
num_list = ""
current_scale = min_scale
while current_scale < max_scale:
check_list += str(current_scale) + " "
num_list += str(pick_time) + " "
current_scale += divide
print(check_list)
print(num_list)
for alpha in [0]: # [0.3, 0.5, 1.0]:
for local_t in [0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=49 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data, local_t, num_list,check_list, local_t, alpha)
write_slurm_sh_faster(
"type49crop_lgq_t_%.3f_alpha_%.2f_divide%d_lr%.4f" % (
local_t, alpha, divide, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==-49:
#only run on pytorch environment, not base environment
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [-0.1,-0.3,-0.5,-1]: # [0.3, 0.5, 1.0]:
for local_t in [0.18]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=49 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type49align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==50:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0, 0.1,0.5,1.0,2.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=50 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type50align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==51:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0, 0.1,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=51 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type51align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==52:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0, 0.1,0.2,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=52 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type52_1v1_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==53:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0, 0.1,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=53 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type53align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==54:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05]:
for alpha in [0, 0.1,0.5,1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.15,0.18,0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=54 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type54align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==55:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=55 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type55align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==551:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=55 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type55align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==550:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0.1]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
for pred_dim in [256,1024,2048]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=55 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 --pred_dim=%d " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha,pred_dim)
write_slurm_sh_faster(
"type55dim%d_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (pred_dim,local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==56:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.05,0.06]:
for alpha in [0, 0.05,0.1,0.2]: # [0.3, 0.5, 1.0]:
for local_t in [0.18, 0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=56 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data,local_t, num_roi,crop_size, local_t, alpha)
write_slurm_sh_faster(
"type56align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (local_t, alpha, num_roi,crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==58:
for learning_rate in [0.06]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=58 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh("multimoco_proj_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name)
elif args.type==59:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=59 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type59_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==60:
for num_roi in [3,6,10,15,20,25,30]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=60 --num_roi=%d " \
" --mask_size=%d --local_t=%.4f --align=1 " % \
(learning_rate, args.data, epoch, 256,
256,learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type60_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==61:
#for num_roi in ['','6']:
# for crop_size in ['','96']:
indicate_list=[['',''],['6','96']]
for indication in indicate_list:
num_roi = indication[0]
crop_size= indication[1]
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=61 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --align=1 " % \
(learning_rate, args.data, epoch, 256, 256,
learning_rate / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type61_lgq_t_%.3f_%s_%s_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==62:
for learning_rate in [0.06]:
for alpha in [0,1.0]:#0 denotes only shuffling to influence
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=62 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh("pixelembedshufflemoco_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name)
elif args.type==63:
for learning_rate in [0.06]:
for alpha in [0,1.0]:#0 denotes only shuffling to influence
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=63 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh("pixelGLsync_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name)
elif args.type == 64:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0,0.1,0.2,0.5, 1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=64 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data, local_t, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type64align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 65:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0,0.1,0.2,0.5, 1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=65 --nmb_crops 1 %d " \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data, local_t, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type65align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 66:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for alpha in [0, 0.1, 0.2, 0.5, 1.0]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=66 --nmb_crops 1 %d " \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data, local_t, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type66align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type == 67:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06,0.08,0.09]:
for alpha in [0, 0.1, 0.2, 0.5]: # [0.3, 0.5, 1.0]:
for local_t in [0.20]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=0.0003 " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f --choose=0,1,2,3,4,5,6,7 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=67 --nmb_crops 1 %d " \
" --size_crops 224 %d --local_t=%.4f --alpha=%.2f --align=1 " % \
(learning_rate, args.data, local_t, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type67align_lgq_t_%.3f_alpha_%.2f_%d_%d_lr%.4f" % (
local_t, alpha, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==68:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=68 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type68_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==69:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=69 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type69_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==70:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=70 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type70_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==71:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for alpha in [0,0.05,0.1,0.2]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=71 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --alpha=%.4f " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t,alpha)
write_slurm_sh_faster(
"type71_lgq_t_%.3f_%d_%d_lr%.4f_alpha%.4f" % (local_t, num_roi, crop_size, learning_rate,alpha),
command_line, queue_name)
time.sleep(1)
elif args.type==72:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=72 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type72_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==73:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=73 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type73_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==74:
for crop_size in [64,96,128,160,192]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=74 --mask_size %d " \
" --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, crop_size, local_t)
write_slurm_sh_faster(
"type74_lgq_t_%.3f_mask%d_lr%.4f" % (local_t, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==75:
for num_roi in [3,6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=75 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type75_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==76 or args.type==98:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(9):
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type,num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type,shuffle_mode,local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==-76:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [0,1,7]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=76 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d --mlp_bn_stat=0 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_faster(
"type76_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (shuffle_mode,local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==77:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [0,1,2,3,5,6]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=77 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_faster(
"type77_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (shuffle_mode,local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==78:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [0,1,3,4,5,7]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=78 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_faster(
"type78_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (shuffle_mode,local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==79:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(2,11):
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=79 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode)
write_slurm_sh_faster(
"type79_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==80:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1,5,7]:
for mlp_bn_stat in [0,1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=80 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" \
" --mlp_bn_stat=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode,mlp_bn_stat)
write_slurm_sh_faster(
"type80_%d_lgq_t_%.3f_%d_%d_lr%.4f_bnmode%d" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate,mlp_bn_stat),
command_line, queue_name)
time.sleep(1)
elif args.type==81:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1,5,7]:
for mlp_bn_stat in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=81 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" \
" --mlp_bn_stat=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode,mlp_bn_stat)
write_slurm_sh_faster(
"type81_%d_lgq_t_%.3f_%d_%d_lr%.4f_bnmode%d" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate,mlp_bn_stat),
command_line, queue_name)
time.sleep(1)
elif args.type==82:
for num_roi in [6,16,32,64]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1,5]:
for mlp_bn_stat in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=82 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" \
" --mlp_bn_stat=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode,mlp_bn_stat)
write_slurm_sh_faster(
"type82_%d_lgq_t_%.3f_%d_%d_lr%.4f_bnmode%d" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate,mlp_bn_stat),
command_line, queue_name)
time.sleep(1)
elif args.type == 83 or args.type==84:
for num_roi in [1,3,5,10]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for alpha in [0.1,0.2,0.5,1.0,2.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --num_roi %d" \
" --mask_size %d --local_t=%.4f --align=1 --alpha=%f " \
" " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t,args.type, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type%d_lgq_t_%.3f_%d_%d_lr%.4f_alpha%f" % (args.type,
local_t, num_roi, crop_size, learning_rate,alpha),
command_line, queue_name)
time.sleep(1)
elif args.type==85:
for num_roi in [6,16,32,64]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1,5]:
for mlp_bn_stat in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=85 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" \
" --mlp_bn_stat=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode,mlp_bn_stat)
write_slurm_sh_faster(
"type85_%d_lgq_t_%.3f_%d_%d_lr%.4f_bnmode%d" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate,mlp_bn_stat),
command_line, queue_name)
time.sleep(1)
elif args.type==86:
for num_roi in [6,16,32]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1,5,7]:
for mlp_bn_stat in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=86 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode)
write_slurm_sh_faster(
"type86_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==87 or args.type==88 or args.type==93 or args.type==94 or args.type==95 or args.type==96:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t,args.type, num_roi, crop_size, local_t)
write_slurm_sh_faster(
"type%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type,
local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==89 or args.type==90:
for num_roi in [1,5,10]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for alpha in [0.1,0.2,0.5,1.0,2.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --num_roi %d" \
" --mask_size %d --local_t=%.4f --align=1 --alpha=%f " \
" " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t,args.type, num_roi, crop_size, local_t, alpha)
write_slurm_sh_faster(
"type%d_lgq_t_%.3f_%d_%d_lr%.4f_alpha%f" % (args.type,
local_t, num_roi, crop_size, learning_rate,alpha),
command_line, queue_name)
time.sleep(1)
elif args.type==91:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t)
write_slurm_sh_faster(
"type%d_lgq_t_%.3f_lr%.4f" % (args.type, local_t, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==92:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(4):
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,shuffle_mode)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_lr%.4f" % (args.type,shuffle_mode, local_t, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==97:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(4):
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=97 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, shuffle_mode)
write_slurm_sh_faster(
"type97_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
shuffle_mode, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==99 or args.type==103 or args.type==104 or args.type==105 \
or args.type==106 or args.type==107 or args.type==108 or args.type==109 \
or args.type==110 or args.type==111 or args.type==112 or args.type==113:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,shuffle_mode)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_lr%.4f" % (args.type,shuffle_mode, local_t, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==126 or args.type==127 or args.type==129 or args.type==131:
for learning_rate in [0.03]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(8):
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,shuffle_mode)
write_slurm_sh_faster(
"type%dablation_%d_lgq_t_%.3f_lr%.4f" % (args.type,shuffle_mode, local_t, learning_rate),
command_line, queue_name,environment=1)
time.sleep(1)
elif args.type==133 or args.type==134:
for learning_rate in [0.03]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in range(3):
for momentum_weight_decay in [0.9,0.99,0.999]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d --use_fp16=1 --momentum_stat=%f" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t, shuffle_mode,momentum_weight_decay)
write_slurm_sh_faster(
"type%dablation_%d_%f_lgq_t_%.3f_lr%.4f" % (
args.type, shuffle_mode,momentum_weight_decay, local_t, learning_rate),
command_line, queue_name, environment=1)
time.sleep(1)
elif args.type==128 or args.type==130 or args.type==132 or args.type==135 or args.type==136:
for learning_rate in [0.03]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,2,4,8,16,32,64,128]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,group_norm_size)
write_slurm_sh_faster(
"type%dgroupablation_%d_lgq_t_%.3f_lr%.4f" % (args.type,group_norm_size, local_t, learning_rate),
command_line, queue_name,environment=1)
time.sleep(1)
elif args.type==152:
for learning_rate in [0.03]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,2,4,8,16,32,64,128]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,group_norm_size)
write_slurm_sh_faster(
"type%dgroup_%d_lgq_t_%.3f_lr%.4f" % (args.type,group_norm_size, local_t, learning_rate),
command_line, queue_name,environment=0)
time.sleep(1)
elif args.type==137 or args.type==138:
for learning_rate in [0.03]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t)
write_slurm_sh_faster(
"type%d2bnablation_lgq_t_%.3f_lr%.4f" % (args.type,local_t, learning_rate),
command_line, queue_name,environment=1)
time.sleep(1)
elif args.type==118:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [1]:
for conv_size in [1,2,3,4]:
for stride_size in [1,2,3]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d --loco_conv_size=%d " \
"--loco_conv_stride=%d" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t, shuffle_mode,conv_size,stride_size)
write_slurm_sh_faster(
"type%d_%d_conv%d_%d_lr%.4f" % (args.type, shuffle_mode, conv_size,
stride_size,learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==114:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,2,4,8]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,group_norm_size)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_lr%.4f" % (args.type,group_norm_size, local_t, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==115 or args.type==116 or args.type==117 or args.type==120 \
or args.type==121 or args.type==122 or args.type==123 or args.type==124:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,8]:
for alpha in [1.0,3.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.14 0.05" \
" --size_crops 224 96 --nmb_crops 2 6 --max_scale_crops 1.0 0.14 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --alpha=%f " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t,group_norm_size,alpha)
write_slurm_sh_faster(
"type%d_%d_alpha%f_lgq_t_%.3f_lr%.4f" % (args.type,group_norm_size,alpha, local_t, learning_rate),
command_line, queue_name,gpu_memory=True)
time.sleep(1)
elif args.type==-120:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1]:
for num_crops in [4,8,16,32]:
same_alpha = int(num_crops / 2) - 1
iter_alpha =[same_alpha,1.0] if same_alpha!=1 else [1.0]
for alpha in iter_alpha:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.14 " \
" --size_crops 96 --nmb_crops %d --max_scale_crops 1.0 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --alpha=%f --use_fp16=1" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_crops,abs(args.type), local_t, group_norm_size, alpha)
write_slurm_sh_faster(
"type%d_%d_%d_alpha%f_lgq_t_%.3f_lr%.4f" % (
args.type,num_crops, group_norm_size, alpha, local_t, learning_rate),
command_line, queue_name, gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==139 or args.type==140 or args.type==141 or args.type==142 \
or args.type==143 or args.type==144 or args.type==145 or args.type==146 or args.type==147:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1]:
for num_crops in [4,8,16]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.14 " \
" --size_crops 96 --nmb_crops %d --max_scale_crops 1.0 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_crops,args.type, local_t, group_norm_size)
write_slurm_sh_faster(
"type%dviewnorm_%d_%d_lgq_t_%.3f_lr%.4f" % (
args.type, num_crops,group_norm_size, local_t, learning_rate),
command_line, queue_name, gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==148 or args.type==149 or args.type==150:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1]:
for num_crops in [4,8,16,32]:
for crop_size in [224,96]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.2 " \
" --size_crops %d --nmb_crops %d --max_scale_crops 1.0 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, crop_size,num_crops, args.type, local_t, group_norm_size)
write_slurm_sh_faster(
"type%dviewnorm_%d_%d_group%d_lgq_t_%.3f_lr%.4f" % (
args.type, num_crops,crop_size, group_norm_size, local_t, learning_rate),
command_line, queue_name, gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==151:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
" --type=%d --min_scale_crops 0.14 0.05 " \
" --size_crops 224 96 --nmb_crops 4 6 --max_scale_crops 1.0 0.14" \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 --alpha 1.0" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type, local_t, group_norm_size)
write_slurm_sh_faster(
"type%dmultiquery_viewkey_group%d_lgq_t_%.3f_lr%.4f" % (
args.type, group_norm_size, local_t, learning_rate),
command_line, queue_name, gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==125:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for momentum_stat in [0.9,0.99,0.999]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.14 0.05" \
" --size_crops 224 96 --nmb_crops 2 6 --max_scale_crops 1.0 0.14 --type=%d " \
" --local_t=%.4f --align=1 --momentum_stat=%f " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,256,
learning_rate * args.node / 100,
local_t, args.type, local_t, momentum_stat)
write_slurm_sh_faster(
"type%d_momentum%f_lgq_t_%.3f_lr%.4f" % (
args.type, momentum_stat, local_t, learning_rate),
command_line, queue_name, gpu_memory=True)
time.sleep(1)
elif args.type==-108:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for batch_size in [1024]:
for shuffle_mode in [1]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * batch_size/256, args.data, epoch, batch_size,
256,
learning_rate * batch_size/256/ 100,
local_t, abs(args.type), local_t,shuffle_mode)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_lr%.4f" % (args.type,shuffle_mode, local_t, learning_rate*batch_size/256),
command_line, queue_name,gpu_memory=True)
time.sleep(1)
elif args.type==100:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,2,4,8]:
command_line = "python3 main_adco.py --mode=1 --lr=%f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --group_norm_size=%d " % \
(learning_rate/2, args.data, epoch, 128,
128,
learning_rate/ 200,
local_t,args.type, num_roi, crop_size, local_t,group_norm_size)
write_slurm_sh_faster(
"type%d_group%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type,group_norm_size,
local_t, num_roi, crop_size, learning_rate),
command_line, queue_name,gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==101:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_num in [1,2,4,8]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=101 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --group_norm_size=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, group_num)
write_slurm_sh_faster(
"type101_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
group_num, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.type==102:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [0,1,7]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, args.type,num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_faster(
"type%d_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type,shuffle_mode,local_t, num_roi, crop_size, learning_rate),
command_line, queue_name)
time.sleep(1)
elif args.mode==2:
if args.type==58:
for learning_rate in [0.06]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=100 " \
"--batch_size=256 --knn_batch_size=256 --cos=1 --lr_final=%.4f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=0.2 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=58 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--alpha=%.4f " \
" " % (learning_rate,args.data,learning_rate/100,alpha)
write_slurm_sh_multi("multimoco_proj_alpha_%.2f_lr_%.4f"%(alpha,learning_rate), command_line, queue_name,
nodes=args.node,gpu_per_node=args.gpu)
elif args.type==59:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [800]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=59 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate*args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate*args.node / 100,
local_t, num_roi, crop_size, local_t)
write_slurm_sh_multi(
"type59_lgq_t_%.3f_%d_%d_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
elif args.type==61:
for num_roi in ['','6']:
for crop_size in ['','96']:
for learning_rate in [0.04,0.06,0.08]:
for local_t in [0.2]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=61 --nmb_crops 1 %s" \
" --size_crops 224 %s --local_t=%.4f --align=1 --ngpu=%d " % \
(learning_rate, args.data, epoch, 256,256,
learning_rate / 100,
local_t, num_roi, crop_size, local_t,args.gpu)
write_slurm_sh_multi(
"type61_lgq_t_%.3f_%s_%s_lr%.4f" % (local_t, num_roi, crop_size, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==77:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for shuffle_mode in [5]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=77 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --shuffle_mode=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t,shuffle_mode)
write_slurm_sh_multi(
"type77_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (shuffle_mode,local_t, num_roi, crop_size, learning_rate*args.node),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==87 or args.type==88 or args.type==94:
if args.type==87:
roi_num_list=[32]
elif args.type==88:
roi_num_list = [6,32]
else:
roi_num_list = [0]
for num_roi in roi_num_list:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [800]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 128,
learning_rate * args.node / 100,
local_t,args.type, num_roi, crop_size, local_t)
if args.queue<=1:
write_slurm_sh_multi2(
"type%d_lgq_t_%.3f_%d_%d_lr%.4f_epoch%d" % (args.type,
local_t, num_roi, crop_size,
learning_rate, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"type%d_lgq_t_%.3f_%d_%d_lr%.4f_epoch%d" % (args.type,
local_t, num_roi, crop_size, learning_rate,epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type == 100:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_norm_size in [1,2,4,8,16]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --group_norm_size=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t,args.type, num_roi, crop_size, local_t,group_norm_size)
if args.node>=4:
command_line += " --warmup_epochs=10 "
if args.queue <= 1:
write_slurm_sh_multi2(
"type%d_group%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type,group_norm_size,
local_t, num_roi, crop_size, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"type%d_group%d_lgq_t_%.3f_%d_%d_lr%.4f" % (args.type, group_norm_size,
local_t, num_roi, crop_size,
learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==101:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [100]:
for group_num in [1,2,4,8,16]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=101 --nmb_crops 1 %d" \
" --size_crops 224 %d --local_t=%.4f --align=1 --group_norm_size=%d " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
args.node * 256,
learning_rate * args.node / 100,
local_t, num_roi, crop_size, local_t, group_num)
if args.node >= 4:
command_line += " --warmup_epochs=10 "
if args.queue <= 1:
write_slurm_sh_multi2(
"type101_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
group_num, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"type101_%d_lgq_t_%.3f_%d_%d_lr%.4f" % (
group_num, local_t, num_roi, crop_size, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==119:
for batch_size in [4096]:
#for crop_size in [96]:
if True:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [800]:
for group_num in [1,8,16,32]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * batch_size / 256, args.data, epoch, batch_size,
256,
learning_rate * batch_size / 256 / 100,
local_t, abs(args.type), local_t,group_num)
command_line += " --warmup_epochs=10 "
write_slurm_sh_multi(
"mocov2bigbatch_type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_num, learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu, gpu_memory=True,
environment=1)
elif args.type==115 or args.type==120:
for batch_size in [2048]:
for learning_rate in [0.045]:
for local_t in [0.2]:
for epoch in [800]:
for group_norm_size in [64]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=10 --tensorboard=1 --min_scale_crops 0.14 0.05" \
" --size_crops 224 96 --nmb_crops 2 6 --max_scale_crops 1.0 0.14 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --alpha=%f --use_fp16=1 " % \
(learning_rate * batch_size/256, args.data, epoch, batch_size,
256,
learning_rate * batch_size/256/ 100,
local_t, args.type, local_t,group_norm_size,alpha)
write_slurm_sh_multi(
"multimoco_type%d_%d_alpha%f_lgq_t_%.3f_lr%.4f" % (args.type,group_norm_size,alpha, local_t, learning_rate),
command_line, queue_name,nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==149:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [1000]:
for group_norm_size in [1]:
for num_crops in [4]:
for crop_size in [224]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --min_scale_crops 0.2 " \
" --size_crops %d --nmb_crops %d --max_scale_crops 1.0 --type=%d " \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 " % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
512,
learning_rate * args.node / 100,
local_t, crop_size,num_crops, args.type, local_t, group_norm_size)
write_slurm_sh_multi2(
"mocov2_%dview_type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, num_crops,group_norm_size, learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu, gpu_memory=False,
environment=0)
time.sleep(1)
elif args.type==151:
for learning_rate in [0.06]:
for local_t in [0.2]:
for epoch in [1000]:
for group_norm_size in [1]:
for alpha in [1.0]:
command_line = "python3 main_adco.py --mode=1 --lr=%.4f --data=%s " \
"--dist_url=tcp://localhost:10031 --epochs=%d " \
"--batch_size=%d --knn_batch_size=%d --cos=1 --lr_final=%.8f " \
"--momentum=0.9 --weight_decay=1e-4 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 " \
"--moco_m=0.999 --moco_k=65536 --moco_t=%.4f " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
" --type=%d --min_scale_crops 0.14 0.05 " \
" --size_crops 224 96 --nmb_crops 4 6 --max_scale_crops 1.0 0.14" \
" --local_t=%.4f --align=1 --group_norm_size=%d --use_fp16=1 --alpha=1.0" % \
(learning_rate * args.node, args.data, epoch, args.node * 256,
512,
learning_rate * args.node / 100,
local_t, args.type, local_t, group_norm_size)
write_slurm_sh_multi(
"type%dmultiquery_viewkey_group%d_lgq_t_%.3f_lr%.4f" % (
args.type, group_norm_size, local_t, learning_rate),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.mode==6:
if args.type==0 or args.type==1 or args.type==2 or args.type==3:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [512]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=0.9 " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d " \
% (
args.type, args.data, epoch, batch_size,local_t, num_roi, crop_size, args.node * 64)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,
local_t, num_roi,
crop_size,
epoch),
command_line, queue_name)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"mocov3type%d_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, local_t, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==4 or args.type==5 or args.type==6:
for num_roi in [1]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for group_norm_size in [1,2,4,8]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d " \
% (args.type, args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,group_norm_size)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_%d_%d_epoch%d" % (args.type,
group_norm_size,
learning_rate,
local_t,
num_roi,
crop_size,
epoch),
command_line, queue_name,gpu_memory=True)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,group_norm_size,learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,learning_rate,local_t, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==7 or args.type==8:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for group_norm_size in [1,2,4,8]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
% (args.type, args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,group_norm_size)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_%d_%d_epoch%d" % (args.type,
group_norm_size,
learning_rate,
local_t,
num_roi,
crop_size,
epoch),
command_line, queue_name,gpu_memory=True,environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,group_norm_size,learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,learning_rate,local_t, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==-7:
combine_choice=[1024,16]#[[1024,16],[2048,32],[4096,64]]
for num_roi in [10]:
for crop_size in [96]:
for learning_rate in [0.3]:
for local_t in [1.0]:
for epoch in [1000]:
for batch_size,group_norm_size in combine_choice:
command_line = "python3 main_adco.py --mode=6 --type=7 --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1.5e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.996 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
% ( args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,group_norm_size)
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size, learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size, learning_rate, local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==-13:
combine_choice=[[4096,1],[4096,64]]#[[1024,16],[2048,32],[4096,64]]
for num_roi in [20]:
for crop_size in [96]:
for learning_rate in [0.3]:
for local_t in [1.0]:
for epoch in [1000]:
for batch_size,group_norm_size in combine_choice:
command_line = "python3 main_adco.py --mode=6 --type=13 --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1.5e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.996 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
% ( args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,group_norm_size)
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size, learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size, learning_rate, local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==9 or args.type==10:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for ema_param in [0.001,0.01,0.1]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --momentum_stat=%f --use_fp16=1 " \
% (args.type, args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,ema_param)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%f_%flgq_t_%.3f_%d_%d_epoch%d" % (args.type,
ema_param,
learning_rate,
local_t,
num_roi,
crop_size,
epoch),
command_line, queue_name,gpu_memory=True,environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,group_norm_size,learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,learning_rate,local_t, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==11:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for ema_param in [0.999]:
for group_norm_size in [1,4,8,16]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --momentum_stat=%f --use_fp16=1 --group_norm_size=%d " \
% (args.type, args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,ema_param,group_norm_size)
if args.node == 1:
write_slurm_sh_faster(
"mocov3type%d_%f_%d_%flgq_t_%.3f_%d_%d_epoch%d" % (args.type,
group_norm_size,
ema_param,
learning_rate,
local_t,
num_roi,
crop_size,
epoch),
command_line, queue_name, gpu_memory=True, environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,ema_param, learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,ema_param, learning_rate, local_t, num_roi,
crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==12:
for num_roi in [6]:
for crop_size in [96]:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for group_norm_size in [8]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 --nmb_crops 1 %d " \
" --size_crops 224 %d --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
% (args.type, args.data, epoch, batch_size, learning_rate,local_t, num_roi,
crop_size, args.node * 64,group_norm_size)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_%d_%d_epoch%d" % (args.type,
group_norm_size,
learning_rate,
local_t,
num_roi,
crop_size,
epoch),
command_line, queue_name,gpu_memory=True,environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (args.type,group_norm_size,learning_rate,
local_t, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=False,environment=0)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_%d_%d_epoch%d" % (
args.type, group_norm_size,learning_rate,local_t, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==13 or args.type==14 or args.type==15:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for group_norm_size in [1,4,8,16]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 " \
" --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
% (args.type, args.data, epoch, batch_size, learning_rate,
local_t, args.node * 64, group_norm_size)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_epoch%d" % (args.type,
group_norm_size,
learning_rate,
local_t,
epoch),
command_line, queue_name, gpu_memory=True, environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, learning_rate,
local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=False, environment=0)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==19:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for group_norm_size in [1,4,8,16,32]:
for key_group_norm_size in [1,4,8,16,32]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 " \
" --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --key_group=%d " \
"--use_fp16=1 " \
% (args.type, args.data, epoch, batch_size, learning_rate,
local_t, args.node * 64, group_norm_size,key_group_norm_size)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%d_%flgq_t_%.3f_epoch%d" % (args.type,
group_norm_size,
key_group_norm_size,
learning_rate,
local_t,
epoch),
command_line, queue_name, gpu_memory=True, environment=1)
else:
if args.queue <= 3:
write_slurm_sh_multi2(
"mocov3type%d_%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, key_group_norm_size,learning_rate,
local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=False, environment=0)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, key_group_norm_size,learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==16:
for learning_rate in [0.9]:
for local_t in [1.0]:
for epoch in [100]:
for batch_size in [1024]:
for crop_size in [4,8,16]:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=1e-6 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f --warmup_epochs=10 " \
" --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=1 --use_fp16=1 " \
"--nmb_crops %d" \
% (args.type, args.data, epoch, batch_size, learning_rate,
local_t, args.node * 64,crop_size )
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_epoch%d" % (args.type,
crop_size,
learning_rate,
local_t,
epoch),
command_line, queue_name, gpu_memory=True, environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, crop_size, learning_rate,
local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=False, environment=0)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, crop_size, learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.type==17 or args.type==18:
warmup_epoch=10
for learning_rate in [1.5e-4]:
for local_t in [0.2]:
for epoch in [100]:
for batch_size in [1024]:
if args.type==18:
group_list = [1,2,4,8,16,32,64,128]
else:
group_list = [1]
for group_norm_size in group_list:
command_line = "python3 main_adco.py --mode=6 --type=%d --data=%s " \
"--epochs=%d --start_epoch=0 --batch_size=%d --lr=%f " \
"--weight_decay=0.1 --dist_url=tcp://localhost:10031 --rank=0 " \
"--multiprocessing_distributed=1 --world_size=1 --moco_dim=256 " \
"--mlp_dim=4096 --moco_m=0.99 --moco_t=%f " \
" --align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 " \
"--knn_batch_size=%d --group_norm_size=%d --use_fp16=1 " \
"--warmup_epochs %d -a vit_small --crop_min 0.08 " \
% (args.type, args.data, epoch, batch_size, learning_rate,
local_t, 256 , group_norm_size,warmup_epoch)
if args.node == 1:
write_slurm_sh_faster("mocov3type%d_%d_%flgq_t_%.3f_epoch%d" % (args.type,
group_norm_size,
learning_rate,
local_t,
epoch),
command_line, queue_name, gpu_memory=True, environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, learning_rate,
local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=False, environment=0)
else:
write_slurm_sh_multi(
"mocov3type%d_%d_%f_lgq_t_%.3f_epoch%d" % (
args.type, group_norm_size, learning_rate, local_t, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True, environment=1)
time.sleep(1)
elif args.mode==7:
if args.type==0 or args.type==1 or args.type==2 or args.type==3 or args.type==4:
for num_roi in [16]:
for crop_size in [96]:
for learning_rate in [0.05]:
for barch_size in [512]:
for epoch in [100]:
command_line = "python3 main_adco.py --mode=7 --type=%d " \
" --data=%s --epochs=%d --start_epoch=0 --batch_size=%d " \
"--lr=%f --weight_decay=1e-4 --dist_url=tcp://localhost:10031 " \
"--rank=0 --multiprocessing_distributed=1 --world_size=1 " \
"--moco_dim=2048 --mlp_dim=512 --nmb_crops 1 %d --size_crops 224 %d " \
"--align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 --knn_batch_size=%d "\
%(args.type,args.data,epoch,barch_size,learning_rate,num_roi,crop_size,max(64*args.node,256))
if args.node==1:
write_slurm_sh_faster("simsiamtype%d_%d_%d_epoch%d" % (args.type, num_roi, crop_size,
epoch),command_line, queue_name,)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"simsiamtype%d_%d_%d_epoch%d" % (args.type, num_roi, crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"simsiamtype%d_%d_%d_epoch%d" % (args.type, num_roi, crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==5 or args.type==6 or args.type==7 or args.type==8 or args.type==9:
for learning_rate in [0.05]:
for barch_size in [512]:
for epoch in [100]:
for group_norm_size in [1, 2, 4, 8,16,32,64]:
command_line = "python3 main_adco.py --mode=7 --type=%d " \
" --data=%s --epochs=%d --start_epoch=0 --batch_size=%d " \
"--lr=%f --weight_decay=1e-4 --dist_url=tcp://localhost:10031 " \
"--rank=0 --multiprocessing_distributed=1 --world_size=1 " \
"--moco_dim=2048 --mlp_dim=512 --group_norm_size=%d " \
"--align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 --knn_batch_size=%d " \
"--use_fp16=1 " \
% (args.type, args.data, epoch, barch_size, learning_rate,group_norm_size,
max(64 * args.node, 256))
if args.node == 1:
write_slurm_sh_faster("simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size,
epoch), command_line, queue_name,
gpu_memory=True,environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True,environment=1)
else:
write_slurm_sh_multi(
"simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True,environment=1)
time.sleep(1)
elif args.type==-6:
for learning_rate in [0.05]:
for barch_size in [256,512]:
for epoch in [800]:
for group_norm_size in [8]:
command_line = "python3 main_adco.py --mode=7 --type=%d " \
" --data=%s --epochs=%d --start_epoch=0 --batch_size=%d " \
"--lr=%f --weight_decay=1e-4 --dist_url=tcp://localhost:10031 " \
"--rank=0 --multiprocessing_distributed=1 --world_size=1 " \
"--moco_dim=2048 --mlp_dim=512 --group_norm_size=%d " \
"--align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 --knn_batch_size=%d " \
"--use_fp16=1 " \
% (abs(args.type), args.data, epoch, barch_size, learning_rate,group_norm_size,
max(64 * args.node, 256))
if args.node == 1:
write_slurm_sh_faster("simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size,
epoch), command_line, queue_name,
gpu_memory=True )
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
else:
write_slurm_sh_multi(
"simsiamtype%d_%d_epoch%d" % (args.type,group_norm_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu)
time.sleep(1)
elif args.type==10:
for learning_rate in [0.05]:
for barch_size in [512]:
for epoch in [100]:
for crop_size in [4, 8,16]:
command_line = "python3 main_adco.py --mode=7 --type=%d " \
" --data=%s --epochs=%d --start_epoch=0 --batch_size=%d " \
"--lr=%f --weight_decay=1e-4 --dist_url=tcp://localhost:10031 " \
"--rank=0 --multiprocessing_distributed=1 --world_size=1 " \
"--moco_dim=2048 --mlp_dim=512 --nmb_crops %d " \
"--align=1 --knn_neighbor=20 --knn_freq=1 --tensorboard=1 --knn_batch_size=%d " \
"--use_fp16=1 " \
% (args.type, args.data, epoch, barch_size, learning_rate,crop_size,
max(64 * args.node, 256))
if args.node == 1:
write_slurm_sh_faster("simsiamtype%d_%d_epoch%d" % (args.type,crop_size,
epoch), command_line, queue_name,
gpu_memory=True,environment=1)
else:
if args.queue <= 1:
write_slurm_sh_multi2(
"simsiamtype%d_%d_epoch%d" % (args.type,crop_size,
epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True,environment=1)
else:
write_slurm_sh_multi(
"simsiamtype%d_%d_epoch%d" % (args.type,crop_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=True,environment=1)
time.sleep(1)
elif args.mode==5:
#run swav baseline
if args.type==0:
if args.F is None:
command_line = "python3 main_adco.py --mode=5 --type=0 --data=%s --epochs 200 --lr=0.6 "\
"--lr_final 0.0006 --batch_size=256 --warmup_epochs 0 --freeze_prototypes_niters 5005 "\
"--queue_length 3840 --epoch_queue_starts 15 --dist_url=tcp://localhost:10031 "\
"--knn_batch_size=256 --cos=1 --momentum=0.9 --weight_decay=1e-6 --world_size=1 "\
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 --moco_k=3000 --moco_t=0.1 "\
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 --dump_path %s"%(args.data,dump_path)
write_slurm_sh("swav_baseline" , command_line, queue_name)
else:
args.F= os.path.abspath(args.F)
command_line = "python3 main_adco.py --mode=5 --type=0 --data=%s --epochs 200 --lr=0.6 " \
"--lr_final 0.0006 --batch_size=256 --warmup_epochs 0 --freeze_prototypes_niters 5005 " \
"--queue_length 3840 --epoch_queue_starts 15 --dist_url=tcp://localhost:10031 " \
"--knn_batch_size=256 --cos=1 --momentum=0.9 --weight_decay=1e-6 --world_size=1 " \
"--rank=0 --multiprocessing_distributed=1 --moco_dim=128 --moco_k=3000 --moco_t=0.1 " \
"--knn_neighbor=20 --knn_freq=1 --tensorboard=1 --nmb_crops 2 6 " \
"--size_crops 224 96 --min_scale_crops 0.14 0.05 --max_scale_crops 1.0 0.14 " \
"--resume=%s --dump_path %s " % (args.data,args.F,dump_path)
resume_name= os.path.split(os.path.abspath(args.F))[1]
write_slurm_sh("swav_baseline_resume%s"%resume_name, command_line, queue_name)
elif args.mode==8:
if args.type==0 or args.type==1:
for epoch in [100]:
for batch_size in [2048]:
for lr_w in [0.2]:
for lr_bias in [0.0048]:
for alpha in [0.51]:
command_line="python3 main.py %s --epochs=%d " \
"--batch-size=%d --learning-rate-weights=%f --learning-rate-biases=%f " \
"--weight-decay=1e-6 --lambd=%f --type=%d --knn_neighbor=20 " \
"--knn_freq=1 --knn_batch_size=%d --tensorboard=1 "%(args.data,epoch,
batch_size,lr_w,lr_bias,alpha,args.type,256 )
if args.node==1:
write_slurm_sh_faster("BTtype%d_%d_epoch%d" % (args.type,batch_size,epoch), command_line, queue_name,
gpu_memory=False, environment=0)
else:
write_slurm_sh_multi2(
"BTtype%d_%d_epoch%d" % (args.type, batch_size, epoch),
command_line, queue_name, nodes=args.node, gpu_per_node=args.gpu,
gpu_memory=False, environment=0)
elif args.type==2:
for epoch in [100]:
for batch_size in [1024]:
for lr_w in [0.2]:
for lr_bias in [0.0048]:
for alpha in [0.51]:
for group_size in [2,4,8,16,32]:
command_line = "python3 main.py %s --epochs=%d " \
"--batch-size=%d --learning-rate-weights=%f --learning-rate-biases=%f " \
"--weight-decay=1e-6 --lambd=%f --type=%d --knn_neighbor=20 " \
"--knn_freq=1 --knn_batch_size=%d --tensorboard=1 --group_norm_size=%d " % (args.data, epoch,
batch_size, lr_w,
lr_bias, alpha,
args.type, 256,group_size)
write_slurm_sh_faster("BTtype%d_%d_%d_epoch%d" % (args.type,group_size, batch_size,epoch), command_line, queue_name,
gpu_memory=False, environment=0)
elif args.mode==0:
#used for finetuning, which will submit finetune jobs and a comment for which
use_bn=args.type
for lr in [20]:
for weight_decay in [1e-6,1e-7,1e-8,1e-9]:
command_line = "python3 lincls.py --data=%s --dist-url=tcp://localhost:10031 " \
"--pretrained='%s' --lr=%.4f --final_lr=%.8f --dataset=ImageNet --use_bn=%d --wd %.8f" % (
args.data, args.F, lr, lr / 100, use_bn,weight_decay)
write_slurm_sh("linear_eval_%s_%.4f_bn%d_wd_%f" % (args.comment, lr, use_bn,weight_decay), command_line, queue_name)
time.sleep(1)
elif args.mode==-2:
use_bn = args.type
#type 3:l2 norm linear
for lr in [1.0]:
for weight_decay in [1e-5,1e-6,1e-7,1e-8,1e-9]:
command_line = "python3 lincls.py --data=%s --dist-url=tcp://localhost:10031 --batch-size=4096 " \
"--pretrained='%s' --lr=%.4f --final_lr=%.8f --dataset=ImageNet --use_bn=%d --wd %.8f" % (
args.data, args.F, lr, lr / 100, use_bn, weight_decay)
write_slurm_sh("linearb4096_eval_%s_%.4f_bn%d_wd_%.8f" % (args.comment, lr, use_bn, weight_decay), command_line,
queue_name)
elif args.mode==-1:
command_line = "python3 encode.py --data=%s --dist-url=tcp://localhost:10031 " \
"--pretrained='%s' --dataset=ImageNet " % (args.data, args.F)
write_slurm_sh("encode_%s" % (args.comment), command_line, queue_name)
elif args.mode==-3:
command_line = "python3 main_adco.py --sym=0 --lr=0.03 --memory_lr=3 --moco_t=0.12 " \
"--mem_t=0.02 --data=%s --dist_url=tcp://localhost:10001 --mode=0 " \
"--epochs=200 --moco_dim=128 --moco_m=0.999 --moco_k=65536 --cluster=65536 " \
"--knn_neighbor=20 --knn_freq=1 --data=imagenet --batch_size=256 --ad_init=1 "%(args.data)
write_slurm_sh("type0",command_line,queue_name)
elif args.mode==-4:
use_bn = args.type
vit_model =True
for lr in [0.05,0.1]:
for weight_decay in [0]:
for model_type in [0]:
command_line ="python lincls_lars.py -a resnet50 --dist-url 'tcp://localhost:10001' " \
"--multiprocessing-distributed --world-size 1 --rank 0 --pretrained='%s' --lr %f --wd %f " \
"--lars --data %s --use_bn=%d --model_type=%d "%(args.F,lr,
weight_decay,args.data,use_bn,model_type)
if vit_model:
command_line +=" --arch vit_small"
write_slurm_sh("linear_larsb4096_eval_%s_bn%d_%.4f_wd_%.8f" % (args.comment, use_bn,lr,weight_decay),
command_line,
queue_name)
elif args.mode==-40:
use_bn = args.type
study_dir = os.path.abspath(args.F)
checkpoint_name = "checkpoint_0099.pth.tar"
for item in os.listdir(study_dir):
if item== checkpoint_name:
current_model_path = os.path.join(study_dir,item)
current_dir = study_dir
current_comment = os.path.split(current_dir)[1]
else:
current_dir = os.path.join(study_dir,item)
current_comment = os.path.split(current_dir)[1]
current_model_path = find_checkpoint(current_dir,checkpoint_name)
if current_model_path is None:
print("%s dir did not find checkpoint"%current_dir)
continue
if not os.path.exists(current_model_path):
print("%s model path did not exist"%current_model_path)
continue
print("fintune %s model"%current_model_path)
for lr in [0.05, 0.1]:
for weight_decay in [0]:
for model_type in [0]:
command_line = "python lincls_lars.py -a resnet50 --dist-url 'tcp://localhost:10001' " \
"--multiprocessing-distributed --world-size 1 --rank 0 --pretrained='%s' --lr %f --wd %f " \
"--lars --data %s --use_bn=%d --model_type=%d " % (current_model_path, lr,
weight_decay, args.data, use_bn,
model_type)
write_slurm_sh(
"linear_larsb4096_eval_%s_bn%d_%.4f_wd_%.8f" % (str(args.comment)+current_comment, use_bn, lr, weight_decay),
command_line,
queue_name)
elif args.mode==-5:
config_dict={}
config_path = os.path.join(os.getcwd(),"detection")
config_path = os.path.join(config_path,"configs")
config_dict['VOC']=os.path.join(config_path,"pascal_voc_R_50_C4_24k_loco.yaml")
config_dict['VOC_freeze'] = os.path.join(config_path, "pascal_voc_R_50_C4_24k_loco_freeze.yaml")
config_dict['COCO'] = os.path.join(config_path,"coco_R_50_C4_2x.yaml_loco.yaml")
config_dict['COCO_freeze'] =os.path.join(config_path,"coco_R_50_C4_2x.yaml_loco_freeze.yaml")
model_path = os.path.abspath(args.F)
model_name = os.path.split(model_path)[1].replace(".pkl","")
for kk in range(5):
for config_now in ['VOC','VOC_freeze']:
command_line = "python detection/train_net.py --config-file %s --num-gpus 8" \
" MODEL.WEIGHTS %s"%(config_dict[config_now],args.F)
write_slurm_sh_faster("detection_%s_run%d_%s" % (config_now, kk,model_name),
command_line, queue_name, gpu_memory=True)
for config_now in ['COCO',"COCO_freeze"]:
command_line = "python detection/train_net.py --config-file %s --num-gpus 8" \
" MODEL.WEIGHTS %s" % (config_dict[config_now], args.F)
write_slurm_sh_faster("detection_%s_%s" % (config_now, model_name),
command_line, queue_name, gpu_memory=True)
elif args.mode==-6:
#finetune with mocov3 protocol
for lr in [0.03,0.06,0.1,0.15,0.12]:
for weight_decay in [0]:
command_line ="python main_lincls.py -a resnet50 --dist-url 'tcp://localhost:10001' " \
"--multiprocessing-distributed --world-size 1 --rank 0 --pretrained='%s' --lr %f --wd %f " \
" %s "%(args.F,lr,weight_decay,args.data)
write_slurm_sh("linear_main_lincls_%s_%.4f_wd_%.8f" % (args.comment, lr,weight_decay),
command_line,
queue_name)
| 72.542391 | 156 | 0.428224 | [
"MIT"
] | wang3702/barlowtwins | run_slurm.py | 266,956 | Python |
# -*- coding: UTF-8 -*-
#
# Copyright 2018 Joachim Lusiardi
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
"""
Implements the Secure Remote Password (SRP) algorithm. More information can be found on
https://tools.ietf.org/html/rfc5054. See HomeKit spec page 36 for adjustments imposed by Apple.
"""
import math
import hashlib
import os
class Srp:
def __init__(self):
# generator as defined by 3072bit group of RFC 5054
self.g = int(b'5', 16)
# modulus as defined by 3072bit group of RFC 5054
self.n = int(b'''\
FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08\
8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B\
302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9\
A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6\
49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8\
FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D\
670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C\
180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718\
3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D\
04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D\
B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226\
1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C\
BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC\
E0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF''', 16)
# HomeKit requires SHA-512 (See page 36)
self.h = hashlib.sha512
self.A = None
self.B = None
self.salt = None
self.username = None
self.password = None
@staticmethod
def generate_private_key():
"""
Static function to generate a 16 byte random key.
:return: the key as an integer
"""
# see
# - https://github.com/jlusiardi/homekit_python/issues/185#issuecomment-616344895 and
# - https://cryptography.io/en/latest/random-numbers/
return int.from_bytes(os.urandom(16), byteorder="big")
def _calculate_k(self) -> int:
# calculate k (see https://tools.ietf.org/html/rfc5054#section-2.5.3)
hash_instance = self.h()
n = Srp.to_byte_array(self.n)
g = bytearray.fromhex((383 * '00' + '05')) # 383 * b'0' + '5'.encode()
hash_instance.update(n)
hash_instance.update(g)
k = int.from_bytes(hash_instance.digest(), "big")
return k
def _calculate_u(self) -> int:
if self.A is None:
raise RuntimeError('Client\'s public key is missing')
if self.B is None:
raise RuntimeError('Server\'s public key is missing')
hash_instance = self.h()
A_b = Srp.to_byte_array(self.A)
B_b = Srp.to_byte_array(self.B)
hash_instance.update(A_b)
hash_instance.update(B_b)
u = int.from_bytes(hash_instance.digest(), "big")
return u
def get_session_key(self) -> int:
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.get_shared_secret()))
hash_value = int.from_bytes(hash_instance.digest(), "big")
return hash_value
@staticmethod
def to_byte_array(num: int) -> bytearray:
return bytearray(num.to_bytes(int(math.ceil(num.bit_length() / 8)), "big"))
def _calculate_x(self) -> int:
i = (self.username + ':' + self.password).encode()
hash_instance = self.h()
hash_instance.update(i)
hash_value = hash_instance.digest()
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.salt))
hash_instance.update(hash_value)
return int.from_bytes(hash_instance.digest(), "big")
def get_shared_secret(self):
raise NotImplementedError()
class SrpClient(Srp):
"""
Implements all functions that are required to simulate an iOS HomeKit controller
"""
def __init__(self, username: str, password: str):
Srp.__init__(self)
self.username = username
self.password = password
self.salt = None
self.a = self.generate_private_key()
self.A = pow(self.g, self.a, self.n)
self.B = None
def set_salt(self, salt):
if isinstance(salt, bytearray) or isinstance(salt, bytes):
self.salt = int.from_bytes(salt, "big")
else:
self.salt = salt
def get_public_key(self):
return pow(self.g, self.a, self.n)
def set_server_public_key(self, B):
if isinstance(B, bytearray) or isinstance(B, bytes):
self.B = int.from_bytes(B, "big")
else:
self.B = B
def get_shared_secret(self):
if self.B is None:
raise RuntimeError('Server\'s public key is missing')
u = self._calculate_u()
x = self._calculate_x()
k = self._calculate_k()
tmp1 = (self.B - (k * pow(self.g, x, self.n)))
tmp2 = (self.a + (u * x)) # % self.n
S = pow(tmp1, tmp2, self.n)
return S
def get_proof(self):
if self.B is None:
raise RuntimeError('Server\'s public key is missing')
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.n))
hN = bytearray(hash_instance.digest())
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.g))
hg = bytearray(hash_instance.digest())
for index in range(0, len(hN)):
hN[index] ^= hg[index]
u = self.username.encode()
hash_instance = self.h()
hash_instance.update(u)
hu = hash_instance.digest()
K = Srp.to_byte_array(self.get_session_key())
hash_instance = self.h()
hash_instance.update(hN)
hash_instance.update(hu)
hash_instance.update(Srp.to_byte_array(self.salt))
hash_instance.update(Srp.to_byte_array(self.A))
hash_instance.update(Srp.to_byte_array(self.B))
hash_instance.update(K)
return int.from_bytes(hash_instance.digest(), "big")
def verify_servers_proof(self, M):
if isinstance(M, bytearray) or isinstance(M, bytes):
tmp = int.from_bytes(M, "big")
else:
tmp = M
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.A))
hash_instance.update(Srp.to_byte_array(self.get_proof()))
hash_instance.update(Srp.to_byte_array(self.get_session_key()))
return tmp == int.from_bytes(hash_instance.digest(), "big")
class SrpServer(Srp):
"""
Implements all functions that are required to simulate an iOS HomeKit accessory
"""
def __init__(self, username, password):
Srp.__init__(self)
self.username = username
self.salt = SrpServer._create_salt()
self.password = password
self.verifier = self._get_verifier()
self.b = self.generate_private_key()
k = self._calculate_k()
g_b = pow(self.g, self.b, self.n)
self.B = (k * self.verifier + g_b) % self.n
self.A = None
@staticmethod
def _create_salt() -> int:
# see
# - https://github.com/jlusiardi/homekit_python/issues/185#issuecomment-616344895 and
# - https://cryptography.io/en/latest/random-numbers/
return int.from_bytes(os.urandom(16), byteorder="big")
def _get_verifier(self) -> int:
hash_value = self._calculate_x()
v = pow(self.g, hash_value, self.n)
return v
def set_client_public_key(self, A):
self.A = A
def get_salt(self):
return self.salt
def get_public_key(self):
k = self._calculate_k()
return (k * self.verifier + pow(self.g, self.b, self.n)) % self.n
def get_shared_secret(self):
if self.A is None:
raise RuntimeError('Client\'s public key is missing')
tmp1 = self.A * pow(self.verifier, self._calculate_u(), self.n)
return pow(tmp1, self.b, self.n)
def verify_clients_proof(self, m) -> bool:
if self.B is None:
raise RuntimeError('Server\'s public key is missing')
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.n))
hN = bytearray(hash_instance.digest())
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.g))
hg = bytearray(hash_instance.digest())
for index in range(0, len(hN)):
hN[index] ^= hg[index]
u = self.username.encode()
hash_instance = self.h()
hash_instance.update(u)
hu = hash_instance.digest()
K = Srp.to_byte_array(self.get_session_key())
hash_instance = self.h()
hash_instance.update(hN)
hash_instance.update(hu)
hash_instance.update(Srp.to_byte_array(self.salt))
hash_instance.update(Srp.to_byte_array(self.A))
hash_instance.update(Srp.to_byte_array(self.B))
hash_instance.update(K)
return m == int.from_bytes(hash_instance.digest(), "big")
def get_proof(self, m) -> int:
hash_instance = self.h()
hash_instance.update(Srp.to_byte_array(self.A))
hash_instance.update(Srp.to_byte_array(m))
hash_instance.update(Srp.to_byte_array(self.get_session_key()))
return int.from_bytes(hash_instance.digest(), "big")
| 34.743772 | 95 | 0.651029 | [
"Apache-2.0"
] | jlusiardi/homekit_client | homekit/crypto/srp.py | 9,763 | Python |