repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
75
19.8k
code_tokens
sequence
docstring
stringlengths
3
17.3k
docstring_tokens
sequence
sha
stringlengths
40
40
url
stringlengths
87
242
partition
stringclasses
1 value
ratt-ru/PyMORESANE
pymoresane/iuwt.py
gpu_iuwt_recomposition
def gpu_iuwt_recomposition(in1, scale_adjust, store_on_gpu, smoothed_array): """ This function calls the a trous algorithm code to recompose the input into a single array. This is the implementation of the isotropic undecimated wavelet transform recomposition for a GPU. INPUTS: in1 (no default): Array containing wavelet coefficients. scale_adjust (no default): Indicates the number of omitted array pages. store_on_gpu (no default): Boolean specifier for whether the decomposition is stored on the gpu or not. OUTPUTS: recomposiiton Array containing the reconstructed array. """ wavelet_filter = (1./16)*np.array([1,4,6,4,1], dtype=np.float32) # Filter-bank for use in the a trous algorithm. wavelet_filter = gpuarray.to_gpu_async(wavelet_filter) # Determines scale with adjustment and creates a zero array on the GPU to store the output,unless smoothed_array # is given. max_scale = in1.shape[0] + scale_adjust if smoothed_array is None: recomposition = gpuarray.zeros([in1.shape[1], in1.shape[2]], np.float32) else: recomposition = gpuarray.to_gpu(smoothed_array.astype(np.float32)) # Determines whether the array is already on the GPU or not. If not, moves it to the GPU. try: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.float32)) except: gpu_in1 = in1 # Creates a working array on the GPU. gpu_tmp = gpuarray.empty_like(recomposition) # Creates and fills an array with the appropriate scale value. gpu_scale = gpuarray.zeros([1], np.int32) gpu_scale += max_scale-1 # Fetches the a trous kernels. gpu_a_trous_row_kernel, gpu_a_trous_col_kernel = gpu_a_trous() grid_rows = int(in1.shape[1]//32) grid_cols = int(in1.shape[2]//32) # The following loops call the a trous algorithm code to recompose the input. The first loop assumes that there are # non-zero wavelet coefficients at scales above scale_adjust, while the second loop completes the recomposition # on the scales less than scale_adjust. for i in range(max_scale-1, scale_adjust-1, -1): gpu_a_trous_row_kernel(recomposition, gpu_tmp, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_a_trous_col_kernel(gpu_tmp, recomposition, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) recomposition = recomposition[:,:] + gpu_in1[i-scale_adjust,:,:] gpu_scale -= 1 if scale_adjust>0: for i in range(scale_adjust-1, -1, -1): gpu_a_trous_row_kernel(recomposition, gpu_tmp, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_a_trous_col_kernel(gpu_tmp, recomposition, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_scale -= 1 # Return values depend on mode. if store_on_gpu: return recomposition else: return recomposition.get()
python
def gpu_iuwt_recomposition(in1, scale_adjust, store_on_gpu, smoothed_array): """ This function calls the a trous algorithm code to recompose the input into a single array. This is the implementation of the isotropic undecimated wavelet transform recomposition for a GPU. INPUTS: in1 (no default): Array containing wavelet coefficients. scale_adjust (no default): Indicates the number of omitted array pages. store_on_gpu (no default): Boolean specifier for whether the decomposition is stored on the gpu or not. OUTPUTS: recomposiiton Array containing the reconstructed array. """ wavelet_filter = (1./16)*np.array([1,4,6,4,1], dtype=np.float32) # Filter-bank for use in the a trous algorithm. wavelet_filter = gpuarray.to_gpu_async(wavelet_filter) # Determines scale with adjustment and creates a zero array on the GPU to store the output,unless smoothed_array # is given. max_scale = in1.shape[0] + scale_adjust if smoothed_array is None: recomposition = gpuarray.zeros([in1.shape[1], in1.shape[2]], np.float32) else: recomposition = gpuarray.to_gpu(smoothed_array.astype(np.float32)) # Determines whether the array is already on the GPU or not. If not, moves it to the GPU. try: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.float32)) except: gpu_in1 = in1 # Creates a working array on the GPU. gpu_tmp = gpuarray.empty_like(recomposition) # Creates and fills an array with the appropriate scale value. gpu_scale = gpuarray.zeros([1], np.int32) gpu_scale += max_scale-1 # Fetches the a trous kernels. gpu_a_trous_row_kernel, gpu_a_trous_col_kernel = gpu_a_trous() grid_rows = int(in1.shape[1]//32) grid_cols = int(in1.shape[2]//32) # The following loops call the a trous algorithm code to recompose the input. The first loop assumes that there are # non-zero wavelet coefficients at scales above scale_adjust, while the second loop completes the recomposition # on the scales less than scale_adjust. for i in range(max_scale-1, scale_adjust-1, -1): gpu_a_trous_row_kernel(recomposition, gpu_tmp, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_a_trous_col_kernel(gpu_tmp, recomposition, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) recomposition = recomposition[:,:] + gpu_in1[i-scale_adjust,:,:] gpu_scale -= 1 if scale_adjust>0: for i in range(scale_adjust-1, -1, -1): gpu_a_trous_row_kernel(recomposition, gpu_tmp, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_a_trous_col_kernel(gpu_tmp, recomposition, wavelet_filter, gpu_scale, block=(32,32,1), grid=(grid_cols, grid_rows)) gpu_scale -= 1 # Return values depend on mode. if store_on_gpu: return recomposition else: return recomposition.get()
[ "def", "gpu_iuwt_recomposition", "(", "in1", ",", "scale_adjust", ",", "store_on_gpu", ",", "smoothed_array", ")", ":", "wavelet_filter", "=", "(", "1.", "/", "16", ")", "*", "np", ".", "array", "(", "[", "1", ",", "4", ",", "6", ",", "4", ",", "1", "]", ",", "dtype", "=", "np", ".", "float32", ")", "wavelet_filter", "=", "gpuarray", ".", "to_gpu_async", "(", "wavelet_filter", ")", "max_scale", "=", "in1", ".", "shape", "[", "0", "]", "+", "scale_adjust", "if", "smoothed_array", "is", "None", ":", "recomposition", "=", "gpuarray", ".", "zeros", "(", "[", "in1", ".", "shape", "[", "1", "]", ",", "in1", ".", "shape", "[", "2", "]", "]", ",", "np", ".", "float32", ")", "else", ":", "recomposition", "=", "gpuarray", ".", "to_gpu", "(", "smoothed_array", ".", "astype", "(", "np", ".", "float32", ")", ")", "try", ":", "gpu_in1", "=", "gpuarray", ".", "to_gpu_async", "(", "in1", ".", "astype", "(", "np", ".", "float32", ")", ")", "except", ":", "gpu_in1", "=", "in1", "gpu_tmp", "=", "gpuarray", ".", "empty_like", "(", "recomposition", ")", "gpu_scale", "=", "gpuarray", ".", "zeros", "(", "[", "1", "]", ",", "np", ".", "int32", ")", "gpu_scale", "+=", "max_scale", "-", "1", "gpu_a_trous_row_kernel", ",", "gpu_a_trous_col_kernel", "=", "gpu_a_trous", "(", ")", "grid_rows", "=", "int", "(", "in1", ".", "shape", "[", "1", "]", "//", "32", ")", "grid_cols", "=", "int", "(", "in1", ".", "shape", "[", "2", "]", "//", "32", ")", "for", "i", "in", "range", "(", "max_scale", "-", "1", ",", "scale_adjust", "-", "1", ",", "-", "1", ")", ":", "gpu_a_trous_row_kernel", "(", "recomposition", ",", "gpu_tmp", ",", "wavelet_filter", ",", "gpu_scale", ",", "block", "=", "(", "32", ",", "32", ",", "1", ")", ",", "grid", "=", "(", "grid_cols", ",", "grid_rows", ")", ")", "gpu_a_trous_col_kernel", "(", "gpu_tmp", ",", "recomposition", ",", "wavelet_filter", ",", "gpu_scale", ",", "block", "=", "(", "32", ",", "32", ",", "1", ")", ",", "grid", "=", "(", "grid_cols", ",", "grid_rows", ")", ")", "recomposition", "=", "recomposition", "[", ":", ",", ":", "]", "+", "gpu_in1", "[", "i", "-", "scale_adjust", ",", ":", ",", ":", "]", "gpu_scale", "-=", "1", "if", "scale_adjust", ">", "0", ":", "for", "i", "in", "range", "(", "scale_adjust", "-", "1", ",", "-", "1", ",", "-", "1", ")", ":", "gpu_a_trous_row_kernel", "(", "recomposition", ",", "gpu_tmp", ",", "wavelet_filter", ",", "gpu_scale", ",", "block", "=", "(", "32", ",", "32", ",", "1", ")", ",", "grid", "=", "(", "grid_cols", ",", "grid_rows", ")", ")", "gpu_a_trous_col_kernel", "(", "gpu_tmp", ",", "recomposition", ",", "wavelet_filter", ",", "gpu_scale", ",", "block", "=", "(", "32", ",", "32", ",", "1", ")", ",", "grid", "=", "(", "grid_cols", ",", "grid_rows", ")", ")", "gpu_scale", "-=", "1", "if", "store_on_gpu", ":", "return", "recomposition", "else", ":", "return", "recomposition", ".", "get", "(", ")" ]
This function calls the a trous algorithm code to recompose the input into a single array. This is the implementation of the isotropic undecimated wavelet transform recomposition for a GPU. INPUTS: in1 (no default): Array containing wavelet coefficients. scale_adjust (no default): Indicates the number of omitted array pages. store_on_gpu (no default): Boolean specifier for whether the decomposition is stored on the gpu or not. OUTPUTS: recomposiiton Array containing the reconstructed array.
[ "This", "function", "calls", "the", "a", "trous", "algorithm", "code", "to", "recompose", "the", "input", "into", "a", "single", "array", ".", "This", "is", "the", "implementation", "of", "the", "isotropic", "undecimated", "wavelet", "transform", "recomposition", "for", "a", "GPU", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt.py#L501-L581
train
marcelcaraciolo/foursquare
examples/django/example/djfoursquare/views.py
unauth
def unauth(request): """ logout and remove all session data """ if check_key(request): api = get_api(request) request.session.clear() logout(request) return HttpResponseRedirect(reverse('main'))
python
def unauth(request): """ logout and remove all session data """ if check_key(request): api = get_api(request) request.session.clear() logout(request) return HttpResponseRedirect(reverse('main'))
[ "def", "unauth", "(", "request", ")", ":", "if", "check_key", "(", "request", ")", ":", "api", "=", "get_api", "(", "request", ")", "request", ".", "session", ".", "clear", "(", ")", "logout", "(", "request", ")", "return", "HttpResponseRedirect", "(", "reverse", "(", "'main'", ")", ")" ]
logout and remove all session data
[ "logout", "and", "remove", "all", "session", "data" ]
a8bda33cc2d61e25aa8df72011246269fd98aa13
https://github.com/marcelcaraciolo/foursquare/blob/a8bda33cc2d61e25aa8df72011246269fd98aa13/examples/django/example/djfoursquare/views.py#L22-L30
train
marcelcaraciolo/foursquare
examples/django/example/djfoursquare/views.py
info
def info(request): """ display some user info to show we have authenticated successfully """ if check_key(request): api = get_api(request) user = api.users(id='self') print dir(user) return render_to_response('djfoursquare/info.html', {'user': user}) else: return HttpResponseRedirect(reverse('main'))
python
def info(request): """ display some user info to show we have authenticated successfully """ if check_key(request): api = get_api(request) user = api.users(id='self') print dir(user) return render_to_response('djfoursquare/info.html', {'user': user}) else: return HttpResponseRedirect(reverse('main'))
[ "def", "info", "(", "request", ")", ":", "if", "check_key", "(", "request", ")", ":", "api", "=", "get_api", "(", "request", ")", "user", "=", "api", ".", "users", "(", "id", "=", "'self'", ")", "print", "dir", "(", "user", ")", "return", "render_to_response", "(", "'djfoursquare/info.html'", ",", "{", "'user'", ":", "user", "}", ")", "else", ":", "return", "HttpResponseRedirect", "(", "reverse", "(", "'main'", ")", ")" ]
display some user info to show we have authenticated successfully
[ "display", "some", "user", "info", "to", "show", "we", "have", "authenticated", "successfully" ]
a8bda33cc2d61e25aa8df72011246269fd98aa13
https://github.com/marcelcaraciolo/foursquare/blob/a8bda33cc2d61e25aa8df72011246269fd98aa13/examples/django/example/djfoursquare/views.py#L33-L43
train
marcelcaraciolo/foursquare
examples/django/example/djfoursquare/views.py
check_key
def check_key(request): """ Check to see if we already have an access_key stored, if we do then we have already gone through OAuth. If not then we haven't and we probably need to. """ try: access_key = request.session.get('oauth_token', None) if not access_key: return False except KeyError: return False return True
python
def check_key(request): """ Check to see if we already have an access_key stored, if we do then we have already gone through OAuth. If not then we haven't and we probably need to. """ try: access_key = request.session.get('oauth_token', None) if not access_key: return False except KeyError: return False return True
[ "def", "check_key", "(", "request", ")", ":", "try", ":", "access_key", "=", "request", ".", "session", ".", "get", "(", "'oauth_token'", ",", "None", ")", "if", "not", "access_key", ":", "return", "False", "except", "KeyError", ":", "return", "False", "return", "True" ]
Check to see if we already have an access_key stored, if we do then we have already gone through OAuth. If not then we haven't and we probably need to.
[ "Check", "to", "see", "if", "we", "already", "have", "an", "access_key", "stored", "if", "we", "do", "then", "we", "have", "already", "gone", "through", "OAuth", ".", "If", "not", "then", "we", "haven", "t", "and", "we", "probably", "need", "to", "." ]
a8bda33cc2d61e25aa8df72011246269fd98aa13
https://github.com/marcelcaraciolo/foursquare/blob/a8bda33cc2d61e25aa8df72011246269fd98aa13/examples/django/example/djfoursquare/views.py#L73-L85
train
praekeltfoundation/seaworthy
seaworthy/stream/_timeout.py
stream_timeout
def stream_timeout(stream, timeout, timeout_msg=None): """ Iterate over items in a streaming response from the Docker client within a timeout. :param ~docker.types.daemon.CancellableStream stream: Stream from the Docker client to consume items from. :param timeout: Timeout value in seconds. :param timeout_msg: Message to raise in the exception when a timeout occurs. """ timed_out = threading.Event() def timeout_func(): timed_out.set() stream.close() timer = threading.Timer(timeout, timeout_func) try: timer.start() for item in stream: yield item # A timeout looks the same as the loop ending. So we need to check a # flag to determine whether a timeout occurred or not. if timed_out.is_set(): raise TimeoutError(timeout_msg) finally: timer.cancel() # Close the stream's underlying response object (if it has one) to # avoid potential socket leaks. # This method seems to have more success at preventing ResourceWarnings # than just stream.close() (should this be improved upstream?) # FIXME: Potential race condition if Timer thread closes the stream at # the same time we do here, but hopefully not with serious side effects if hasattr(stream, '_response'): stream._response.close()
python
def stream_timeout(stream, timeout, timeout_msg=None): """ Iterate over items in a streaming response from the Docker client within a timeout. :param ~docker.types.daemon.CancellableStream stream: Stream from the Docker client to consume items from. :param timeout: Timeout value in seconds. :param timeout_msg: Message to raise in the exception when a timeout occurs. """ timed_out = threading.Event() def timeout_func(): timed_out.set() stream.close() timer = threading.Timer(timeout, timeout_func) try: timer.start() for item in stream: yield item # A timeout looks the same as the loop ending. So we need to check a # flag to determine whether a timeout occurred or not. if timed_out.is_set(): raise TimeoutError(timeout_msg) finally: timer.cancel() # Close the stream's underlying response object (if it has one) to # avoid potential socket leaks. # This method seems to have more success at preventing ResourceWarnings # than just stream.close() (should this be improved upstream?) # FIXME: Potential race condition if Timer thread closes the stream at # the same time we do here, but hopefully not with serious side effects if hasattr(stream, '_response'): stream._response.close()
[ "def", "stream_timeout", "(", "stream", ",", "timeout", ",", "timeout_msg", "=", "None", ")", ":", "timed_out", "=", "threading", ".", "Event", "(", ")", "def", "timeout_func", "(", ")", ":", "timed_out", ".", "set", "(", ")", "stream", ".", "close", "(", ")", "timer", "=", "threading", ".", "Timer", "(", "timeout", ",", "timeout_func", ")", "try", ":", "timer", ".", "start", "(", ")", "for", "item", "in", "stream", ":", "yield", "item", "if", "timed_out", ".", "is_set", "(", ")", ":", "raise", "TimeoutError", "(", "timeout_msg", ")", "finally", ":", "timer", ".", "cancel", "(", ")", "if", "hasattr", "(", "stream", ",", "'_response'", ")", ":", "stream", ".", "_response", ".", "close", "(", ")" ]
Iterate over items in a streaming response from the Docker client within a timeout. :param ~docker.types.daemon.CancellableStream stream: Stream from the Docker client to consume items from. :param timeout: Timeout value in seconds. :param timeout_msg: Message to raise in the exception when a timeout occurs.
[ "Iterate", "over", "items", "in", "a", "streaming", "response", "from", "the", "Docker", "client", "within", "a", "timeout", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/stream/_timeout.py#L4-L41
train
tuomas2/automate
src/automate/callable.py
AbstractCallable.get_state
def get_state(self, caller): """ Get per-program state. """ if caller in self.state: return self.state[caller] else: rv = self.state[caller] = DictObject() return rv
python
def get_state(self, caller): """ Get per-program state. """ if caller in self.state: return self.state[caller] else: rv = self.state[caller] = DictObject() return rv
[ "def", "get_state", "(", "self", ",", "caller", ")", ":", "if", "caller", "in", "self", ".", "state", ":", "return", "self", ".", "state", "[", "caller", "]", "else", ":", "rv", "=", "self", ".", "state", "[", "caller", "]", "=", "DictObject", "(", ")", "return", "rv" ]
Get per-program state.
[ "Get", "per", "-", "program", "state", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/callable.py#L74-L83
train
tuomas2/automate
src/automate/callable.py
AbstractCallable.name_to_system_object
def name_to_system_object(self, value): """ Return object for given name registered in System namespace. """ if not self.system: raise SystemNotReady if isinstance(value, (str, Object)): rv = self.system.name_to_system_object(value) return rv if rv else value else: return value
python
def name_to_system_object(self, value): """ Return object for given name registered in System namespace. """ if not self.system: raise SystemNotReady if isinstance(value, (str, Object)): rv = self.system.name_to_system_object(value) return rv if rv else value else: return value
[ "def", "name_to_system_object", "(", "self", ",", "value", ")", ":", "if", "not", "self", ".", "system", ":", "raise", "SystemNotReady", "if", "isinstance", "(", "value", ",", "(", "str", ",", "Object", ")", ")", ":", "rv", "=", "self", ".", "system", ".", "name_to_system_object", "(", "value", ")", "return", "rv", "if", "rv", "else", "value", "else", ":", "return", "value" ]
Return object for given name registered in System namespace.
[ "Return", "object", "for", "given", "name", "registered", "in", "System", "namespace", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/callable.py#L250-L261
train
tuomas2/automate
src/automate/callable.py
AbstractCallable.cancel
def cancel(self, caller): """ Recursively cancel all threaded background processes of this Callable. This is called automatically for actions if program deactivates. """ for o in {i for i in self.children if isinstance(i, AbstractCallable)}: o.cancel(caller)
python
def cancel(self, caller): """ Recursively cancel all threaded background processes of this Callable. This is called automatically for actions if program deactivates. """ for o in {i for i in self.children if isinstance(i, AbstractCallable)}: o.cancel(caller)
[ "def", "cancel", "(", "self", ",", "caller", ")", ":", "for", "o", "in", "{", "i", "for", "i", "in", "self", ".", "children", "if", "isinstance", "(", "i", ",", "AbstractCallable", ")", "}", ":", "o", ".", "cancel", "(", "caller", ")" ]
Recursively cancel all threaded background processes of this Callable. This is called automatically for actions if program deactivates.
[ "Recursively", "cancel", "all", "threaded", "background", "processes", "of", "this", "Callable", ".", "This", "is", "called", "automatically", "for", "actions", "if", "program", "deactivates", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/callable.py#L311-L317
train
tuomas2/automate
src/automate/callable.py
AbstractCallable.give_str
def give_str(self): """ Give string representation of the callable. """ args = self._args[:] kwargs = self._kwargs return self._give_str(args, kwargs)
python
def give_str(self): """ Give string representation of the callable. """ args = self._args[:] kwargs = self._kwargs return self._give_str(args, kwargs)
[ "def", "give_str", "(", "self", ")", ":", "args", "=", "self", ".", "_args", "[", ":", "]", "kwargs", "=", "self", ".", "_kwargs", "return", "self", ".", "_give_str", "(", "args", ",", "kwargs", ")" ]
Give string representation of the callable.
[ "Give", "string", "representation", "of", "the", "callable", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/callable.py#L333-L339
train
fauskanger/mypolr
mypolr/polr_api.py
PolrApi._make_request
def _make_request(self, endpoint, params): """ Prepares the request and catches common errors and returns tuple of data and the request response. Read more about error codes: https://docs.polrproject.org/en/latest/developer-guide/api/#http-error-codes :param endpoint: full endpoint url :type endpoint: str :param params: parameters for the given endpoint :type params: dict :return: Tuple of response data, and the response instance :rtype: dict, requests.Response """ # params = { # **self._base_params, # Mind order to allow params to overwrite base params # **params # } full_params = self._base_params.copy() full_params.update(params) try: r = requests.get(endpoint, full_params) data = r.json() if r.status_code == 401 and not endpoint.endswith('lookup'): raise exceptions.UnauthorizedKeyError elif r.status_code == 400 and not endpoint.endswith('shorten'): raise exceptions.BadApiRequest elif r.status_code == 500: raise exceptions.ServerOrConnectionError return data, r except ValueError as e: raise exceptions.BadApiResponse(e) except requests.RequestException: raise exceptions.ServerOrConnectionError
python
def _make_request(self, endpoint, params): """ Prepares the request and catches common errors and returns tuple of data and the request response. Read more about error codes: https://docs.polrproject.org/en/latest/developer-guide/api/#http-error-codes :param endpoint: full endpoint url :type endpoint: str :param params: parameters for the given endpoint :type params: dict :return: Tuple of response data, and the response instance :rtype: dict, requests.Response """ # params = { # **self._base_params, # Mind order to allow params to overwrite base params # **params # } full_params = self._base_params.copy() full_params.update(params) try: r = requests.get(endpoint, full_params) data = r.json() if r.status_code == 401 and not endpoint.endswith('lookup'): raise exceptions.UnauthorizedKeyError elif r.status_code == 400 and not endpoint.endswith('shorten'): raise exceptions.BadApiRequest elif r.status_code == 500: raise exceptions.ServerOrConnectionError return data, r except ValueError as e: raise exceptions.BadApiResponse(e) except requests.RequestException: raise exceptions.ServerOrConnectionError
[ "def", "_make_request", "(", "self", ",", "endpoint", ",", "params", ")", ":", "full_params", "=", "self", ".", "_base_params", ".", "copy", "(", ")", "full_params", ".", "update", "(", "params", ")", "try", ":", "r", "=", "requests", ".", "get", "(", "endpoint", ",", "full_params", ")", "data", "=", "r", ".", "json", "(", ")", "if", "r", ".", "status_code", "==", "401", "and", "not", "endpoint", ".", "endswith", "(", "'lookup'", ")", ":", "raise", "exceptions", ".", "UnauthorizedKeyError", "elif", "r", ".", "status_code", "==", "400", "and", "not", "endpoint", ".", "endswith", "(", "'shorten'", ")", ":", "raise", "exceptions", ".", "BadApiRequest", "elif", "r", ".", "status_code", "==", "500", ":", "raise", "exceptions", ".", "ServerOrConnectionError", "return", "data", ",", "r", "except", "ValueError", "as", "e", ":", "raise", "exceptions", ".", "BadApiResponse", "(", "e", ")", "except", "requests", ".", "RequestException", ":", "raise", "exceptions", ".", "ServerOrConnectionError" ]
Prepares the request and catches common errors and returns tuple of data and the request response. Read more about error codes: https://docs.polrproject.org/en/latest/developer-guide/api/#http-error-codes :param endpoint: full endpoint url :type endpoint: str :param params: parameters for the given endpoint :type params: dict :return: Tuple of response data, and the response instance :rtype: dict, requests.Response
[ "Prepares", "the", "request", "and", "catches", "common", "errors", "and", "returns", "tuple", "of", "data", "and", "the", "request", "response", "." ]
46eb4fc5ba0f65412634a37e30e05de79fc9db4c
https://github.com/fauskanger/mypolr/blob/46eb4fc5ba0f65412634a37e30e05de79fc9db4c/mypolr/polr_api.py#L42-L74
train
fauskanger/mypolr
mypolr/polr_api.py
PolrApi.shorten
def shorten(self, long_url, custom_ending=None, is_secret=False): """ Creates a short url if valid :param str long_url: The url to shorten. :param custom_ending: The custom url to create if available. :type custom_ending: str or None :param bool is_secret: if not public, it's secret :return: a short link :rtype: str """ params = { 'url': long_url, 'is_secret': 'true' if is_secret else 'false', 'custom_ending': custom_ending } data, r = self._make_request(self.api_shorten_endpoint, params) if r.status_code == 400: if custom_ending is not None: raise exceptions.CustomEndingUnavailable(custom_ending) raise exceptions.BadApiRequest elif r.status_code == 403: raise exceptions.QuotaExceededError action = data.get('action') short_url = data.get('result') if action == 'shorten' and short_url is not None: return short_url raise exceptions.DebugTempWarning
python
def shorten(self, long_url, custom_ending=None, is_secret=False): """ Creates a short url if valid :param str long_url: The url to shorten. :param custom_ending: The custom url to create if available. :type custom_ending: str or None :param bool is_secret: if not public, it's secret :return: a short link :rtype: str """ params = { 'url': long_url, 'is_secret': 'true' if is_secret else 'false', 'custom_ending': custom_ending } data, r = self._make_request(self.api_shorten_endpoint, params) if r.status_code == 400: if custom_ending is not None: raise exceptions.CustomEndingUnavailable(custom_ending) raise exceptions.BadApiRequest elif r.status_code == 403: raise exceptions.QuotaExceededError action = data.get('action') short_url = data.get('result') if action == 'shorten' and short_url is not None: return short_url raise exceptions.DebugTempWarning
[ "def", "shorten", "(", "self", ",", "long_url", ",", "custom_ending", "=", "None", ",", "is_secret", "=", "False", ")", ":", "params", "=", "{", "'url'", ":", "long_url", ",", "'is_secret'", ":", "'true'", "if", "is_secret", "else", "'false'", ",", "'custom_ending'", ":", "custom_ending", "}", "data", ",", "r", "=", "self", ".", "_make_request", "(", "self", ".", "api_shorten_endpoint", ",", "params", ")", "if", "r", ".", "status_code", "==", "400", ":", "if", "custom_ending", "is", "not", "None", ":", "raise", "exceptions", ".", "CustomEndingUnavailable", "(", "custom_ending", ")", "raise", "exceptions", ".", "BadApiRequest", "elif", "r", ".", "status_code", "==", "403", ":", "raise", "exceptions", ".", "QuotaExceededError", "action", "=", "data", ".", "get", "(", "'action'", ")", "short_url", "=", "data", ".", "get", "(", "'result'", ")", "if", "action", "==", "'shorten'", "and", "short_url", "is", "not", "None", ":", "return", "short_url", "raise", "exceptions", ".", "DebugTempWarning" ]
Creates a short url if valid :param str long_url: The url to shorten. :param custom_ending: The custom url to create if available. :type custom_ending: str or None :param bool is_secret: if not public, it's secret :return: a short link :rtype: str
[ "Creates", "a", "short", "url", "if", "valid" ]
46eb4fc5ba0f65412634a37e30e05de79fc9db4c
https://github.com/fauskanger/mypolr/blob/46eb4fc5ba0f65412634a37e30e05de79fc9db4c/mypolr/polr_api.py#L76-L103
train
fauskanger/mypolr
mypolr/polr_api.py
PolrApi._get_ending
def _get_ending(self, lookup_url): """ Returns the short url ending from a short url or an short url ending. Example: - Given `<your Polr server>/5N3f8`, return `5N3f8`. - Given `5N3f8`, return `5N3f8`. :param lookup_url: A short url or short url ending :type lookup_url: str :return: The url ending :rtype: str """ if lookup_url.startswith(self.api_server): return lookup_url[len(self.api_server) + 1:] return lookup_url
python
def _get_ending(self, lookup_url): """ Returns the short url ending from a short url or an short url ending. Example: - Given `<your Polr server>/5N3f8`, return `5N3f8`. - Given `5N3f8`, return `5N3f8`. :param lookup_url: A short url or short url ending :type lookup_url: str :return: The url ending :rtype: str """ if lookup_url.startswith(self.api_server): return lookup_url[len(self.api_server) + 1:] return lookup_url
[ "def", "_get_ending", "(", "self", ",", "lookup_url", ")", ":", "if", "lookup_url", ".", "startswith", "(", "self", ".", "api_server", ")", ":", "return", "lookup_url", "[", "len", "(", "self", ".", "api_server", ")", "+", "1", ":", "]", "return", "lookup_url" ]
Returns the short url ending from a short url or an short url ending. Example: - Given `<your Polr server>/5N3f8`, return `5N3f8`. - Given `5N3f8`, return `5N3f8`. :param lookup_url: A short url or short url ending :type lookup_url: str :return: The url ending :rtype: str
[ "Returns", "the", "short", "url", "ending", "from", "a", "short", "url", "or", "an", "short", "url", "ending", "." ]
46eb4fc5ba0f65412634a37e30e05de79fc9db4c
https://github.com/fauskanger/mypolr/blob/46eb4fc5ba0f65412634a37e30e05de79fc9db4c/mypolr/polr_api.py#L105-L120
train
fauskanger/mypolr
mypolr/polr_api.py
PolrApi.lookup
def lookup(self, lookup_url, url_key=None): """ Looks up the url_ending to obtain information about the short url. If it exists, the API will return a dictionary with information, including the long_url that is the destination of the given short url URL. The lookup object looks like something like this: .. code-block:: python { 'clicks': 42, 'created_at': { 'date': '2017-12-03 00:40:45.000000', 'timezone': 'UTC', 'timezone_type': 3 }, 'long_url': 'https://stackoverflow.com/questions/tagged/python', 'updated_at': { 'date': '2017-12-24 13:37:00.000000', 'timezone': 'UTC', 'timezone_type': 3 } } :param str lookup_url: An url ending or full short url address :param url_key: optional URL ending key for lookups against secret URLs :type url_key: str or None :return: Lookup dictionary containing, among others things, the long url; or None if not existing :rtype: dict or None """ url_ending = self._get_ending(lookup_url) params = { 'url_ending': url_ending, 'url_key': url_key } data, r = self._make_request(self.api_lookup_endpoint, params) if r.status_code == 401: if url_key is not None: raise exceptions.UnauthorizedKeyError('given url_key is not valid for secret lookup.') raise exceptions.UnauthorizedKeyError elif r.status_code == 404: return False # no url found in lookup action = data.get('action') full_url = data.get('result') if action == 'lookup' and full_url is not None: return full_url raise exceptions.DebugTempWarning
python
def lookup(self, lookup_url, url_key=None): """ Looks up the url_ending to obtain information about the short url. If it exists, the API will return a dictionary with information, including the long_url that is the destination of the given short url URL. The lookup object looks like something like this: .. code-block:: python { 'clicks': 42, 'created_at': { 'date': '2017-12-03 00:40:45.000000', 'timezone': 'UTC', 'timezone_type': 3 }, 'long_url': 'https://stackoverflow.com/questions/tagged/python', 'updated_at': { 'date': '2017-12-24 13:37:00.000000', 'timezone': 'UTC', 'timezone_type': 3 } } :param str lookup_url: An url ending or full short url address :param url_key: optional URL ending key for lookups against secret URLs :type url_key: str or None :return: Lookup dictionary containing, among others things, the long url; or None if not existing :rtype: dict or None """ url_ending = self._get_ending(lookup_url) params = { 'url_ending': url_ending, 'url_key': url_key } data, r = self._make_request(self.api_lookup_endpoint, params) if r.status_code == 401: if url_key is not None: raise exceptions.UnauthorizedKeyError('given url_key is not valid for secret lookup.') raise exceptions.UnauthorizedKeyError elif r.status_code == 404: return False # no url found in lookup action = data.get('action') full_url = data.get('result') if action == 'lookup' and full_url is not None: return full_url raise exceptions.DebugTempWarning
[ "def", "lookup", "(", "self", ",", "lookup_url", ",", "url_key", "=", "None", ")", ":", "url_ending", "=", "self", ".", "_get_ending", "(", "lookup_url", ")", "params", "=", "{", "'url_ending'", ":", "url_ending", ",", "'url_key'", ":", "url_key", "}", "data", ",", "r", "=", "self", ".", "_make_request", "(", "self", ".", "api_lookup_endpoint", ",", "params", ")", "if", "r", ".", "status_code", "==", "401", ":", "if", "url_key", "is", "not", "None", ":", "raise", "exceptions", ".", "UnauthorizedKeyError", "(", "'given url_key is not valid for secret lookup.'", ")", "raise", "exceptions", ".", "UnauthorizedKeyError", "elif", "r", ".", "status_code", "==", "404", ":", "return", "False", "action", "=", "data", ".", "get", "(", "'action'", ")", "full_url", "=", "data", ".", "get", "(", "'result'", ")", "if", "action", "==", "'lookup'", "and", "full_url", "is", "not", "None", ":", "return", "full_url", "raise", "exceptions", ".", "DebugTempWarning" ]
Looks up the url_ending to obtain information about the short url. If it exists, the API will return a dictionary with information, including the long_url that is the destination of the given short url URL. The lookup object looks like something like this: .. code-block:: python { 'clicks': 42, 'created_at': { 'date': '2017-12-03 00:40:45.000000', 'timezone': 'UTC', 'timezone_type': 3 }, 'long_url': 'https://stackoverflow.com/questions/tagged/python', 'updated_at': { 'date': '2017-12-24 13:37:00.000000', 'timezone': 'UTC', 'timezone_type': 3 } } :param str lookup_url: An url ending or full short url address :param url_key: optional URL ending key for lookups against secret URLs :type url_key: str or None :return: Lookup dictionary containing, among others things, the long url; or None if not existing :rtype: dict or None
[ "Looks", "up", "the", "url_ending", "to", "obtain", "information", "about", "the", "short", "url", "." ]
46eb4fc5ba0f65412634a37e30e05de79fc9db4c
https://github.com/fauskanger/mypolr/blob/46eb4fc5ba0f65412634a37e30e05de79fc9db4c/mypolr/polr_api.py#L122-L173
train
fauskanger/mypolr
mypolr/cli.py
make_argparser
def make_argparser(): """ Setup argparse arguments. :return: The parser which :class:`MypolrCli` expects parsed arguments from. :rtype: argparse.ArgumentParser """ parser = argparse.ArgumentParser(prog='mypolr', description="Interacts with the Polr Project's API.\n\n" "User Guide and documentation: https://mypolr.readthedocs.io", formatter_class=argparse.ArgumentDefaultsHelpFormatter, epilog="NOTE: if configurations are saved, they are stored as plain text on disk, " "and can be read by anyone with access to the file.") parser.add_argument("-v", "--version", action="store_true", help="Print version and exit.") parser.add_argument("url", nargs='?', default=None, help="The url to process.") api_group = parser.add_argument_group('API server arguments', 'Use these for configure the API. Can be stored locally with --save.') api_group.add_argument("-s", "--server", default=None, help="Server hosting the API.") api_group.add_argument("-k", "--key", default=None, help="API_KEY to authenticate against server.") api_group.add_argument("--api-root", default=DEFAULT_API_ROOT, help="API endpoint root.") option_group = parser.add_argument_group('Action options', 'Configure the API action to use.') option_group.add_argument("-c", "--custom", default=None, help="Custom short url ending.") option_group.add_argument("--secret", action="store_true", help="Set option if using secret url.") option_group.add_argument("-l", "--lookup", action="store_true", help="Perform lookup action instead of shorten action.") manage_group = parser.add_argument_group('Manage credentials', 'Use these to save, delete or update SERVER, KEY and/or ' 'API_ROOT locally in ~/.mypolr/config.ini.') manage_group.add_argument("--save", action="store_true", help="Save configuration (including credentials) in plaintext(!).") manage_group.add_argument("--clear", action="store_true", help="Clear configuration.") return parser
python
def make_argparser(): """ Setup argparse arguments. :return: The parser which :class:`MypolrCli` expects parsed arguments from. :rtype: argparse.ArgumentParser """ parser = argparse.ArgumentParser(prog='mypolr', description="Interacts with the Polr Project's API.\n\n" "User Guide and documentation: https://mypolr.readthedocs.io", formatter_class=argparse.ArgumentDefaultsHelpFormatter, epilog="NOTE: if configurations are saved, they are stored as plain text on disk, " "and can be read by anyone with access to the file.") parser.add_argument("-v", "--version", action="store_true", help="Print version and exit.") parser.add_argument("url", nargs='?', default=None, help="The url to process.") api_group = parser.add_argument_group('API server arguments', 'Use these for configure the API. Can be stored locally with --save.') api_group.add_argument("-s", "--server", default=None, help="Server hosting the API.") api_group.add_argument("-k", "--key", default=None, help="API_KEY to authenticate against server.") api_group.add_argument("--api-root", default=DEFAULT_API_ROOT, help="API endpoint root.") option_group = parser.add_argument_group('Action options', 'Configure the API action to use.') option_group.add_argument("-c", "--custom", default=None, help="Custom short url ending.") option_group.add_argument("--secret", action="store_true", help="Set option if using secret url.") option_group.add_argument("-l", "--lookup", action="store_true", help="Perform lookup action instead of shorten action.") manage_group = parser.add_argument_group('Manage credentials', 'Use these to save, delete or update SERVER, KEY and/or ' 'API_ROOT locally in ~/.mypolr/config.ini.') manage_group.add_argument("--save", action="store_true", help="Save configuration (including credentials) in plaintext(!).") manage_group.add_argument("--clear", action="store_true", help="Clear configuration.") return parser
[ "def", "make_argparser", "(", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", "prog", "=", "'mypolr'", ",", "description", "=", "\"Interacts with the Polr Project's API.\\n\\n\"", "\"User Guide and documentation: https://mypolr.readthedocs.io\"", ",", "formatter_class", "=", "argparse", ".", "ArgumentDefaultsHelpFormatter", ",", "epilog", "=", "\"NOTE: if configurations are saved, they are stored as plain text on disk, \"", "\"and can be read by anyone with access to the file.\"", ")", "parser", ".", "add_argument", "(", "\"-v\"", ",", "\"--version\"", ",", "action", "=", "\"store_true\"", ",", "help", "=", "\"Print version and exit.\"", ")", "parser", ".", "add_argument", "(", "\"url\"", ",", "nargs", "=", "'?'", ",", "default", "=", "None", ",", "help", "=", "\"The url to process.\"", ")", "api_group", "=", "parser", ".", "add_argument_group", "(", "'API server arguments'", ",", "'Use these for configure the API. Can be stored locally with --save.'", ")", "api_group", ".", "add_argument", "(", "\"-s\"", ",", "\"--server\"", ",", "default", "=", "None", ",", "help", "=", "\"Server hosting the API.\"", ")", "api_group", ".", "add_argument", "(", "\"-k\"", ",", "\"--key\"", ",", "default", "=", "None", ",", "help", "=", "\"API_KEY to authenticate against server.\"", ")", "api_group", ".", "add_argument", "(", "\"--api-root\"", ",", "default", "=", "DEFAULT_API_ROOT", ",", "help", "=", "\"API endpoint root.\"", ")", "option_group", "=", "parser", ".", "add_argument_group", "(", "'Action options'", ",", "'Configure the API action to use.'", ")", "option_group", ".", "add_argument", "(", "\"-c\"", ",", "\"--custom\"", ",", "default", "=", "None", ",", "help", "=", "\"Custom short url ending.\"", ")", "option_group", ".", "add_argument", "(", "\"--secret\"", ",", "action", "=", "\"store_true\"", ",", "help", "=", "\"Set option if using secret url.\"", ")", "option_group", ".", "add_argument", "(", "\"-l\"", ",", "\"--lookup\"", ",", "action", "=", "\"store_true\"", ",", "help", "=", "\"Perform lookup action instead of shorten action.\"", ")", "manage_group", "=", "parser", ".", "add_argument_group", "(", "'Manage credentials'", ",", "'Use these to save, delete or update SERVER, KEY and/or '", "'API_ROOT locally in ~/.mypolr/config.ini.'", ")", "manage_group", ".", "add_argument", "(", "\"--save\"", ",", "action", "=", "\"store_true\"", ",", "help", "=", "\"Save configuration (including credentials) in plaintext(!).\"", ")", "manage_group", ".", "add_argument", "(", "\"--clear\"", ",", "action", "=", "\"store_true\"", ",", "help", "=", "\"Clear configuration.\"", ")", "return", "parser" ]
Setup argparse arguments. :return: The parser which :class:`MypolrCli` expects parsed arguments from. :rtype: argparse.ArgumentParser
[ "Setup", "argparse", "arguments", "." ]
46eb4fc5ba0f65412634a37e30e05de79fc9db4c
https://github.com/fauskanger/mypolr/blob/46eb4fc5ba0f65412634a37e30e05de79fc9db4c/mypolr/cli.py#L18-L61
train
ratt-ru/PyMORESANE
pymoresane/iuwt_toolbox.py
estimate_threshold
def estimate_threshold(in1, edge_excl=0, int_excl=0): """ This function estimates the noise using the MAD estimator. INPUTS: in1 (no default): The array from which the noise is estimated OUTPUTS: out1 An array of per-scale noise estimates. """ out1 = np.empty([in1.shape[0]]) mid = in1.shape[1]/2 if (edge_excl!=0) | (int_excl!=0): if edge_excl!=0: mask = np.zeros([in1.shape[1], in1.shape[2]]) mask[edge_excl:-edge_excl, edge_excl:-edge_excl] = 1 else: mask = np.ones([in1.shape[1], in1.shape[2]]) if int_excl!=0: mask[mid-int_excl:mid+int_excl, mid-int_excl:mid+int_excl] = 0 else: mask = np.ones([in1.shape[1], in1.shape[2]]) for i in range(in1.shape[0]): out1[i] = np.median(np.abs(in1[i,mask==1]))/0.6745 return out1
python
def estimate_threshold(in1, edge_excl=0, int_excl=0): """ This function estimates the noise using the MAD estimator. INPUTS: in1 (no default): The array from which the noise is estimated OUTPUTS: out1 An array of per-scale noise estimates. """ out1 = np.empty([in1.shape[0]]) mid = in1.shape[1]/2 if (edge_excl!=0) | (int_excl!=0): if edge_excl!=0: mask = np.zeros([in1.shape[1], in1.shape[2]]) mask[edge_excl:-edge_excl, edge_excl:-edge_excl] = 1 else: mask = np.ones([in1.shape[1], in1.shape[2]]) if int_excl!=0: mask[mid-int_excl:mid+int_excl, mid-int_excl:mid+int_excl] = 0 else: mask = np.ones([in1.shape[1], in1.shape[2]]) for i in range(in1.shape[0]): out1[i] = np.median(np.abs(in1[i,mask==1]))/0.6745 return out1
[ "def", "estimate_threshold", "(", "in1", ",", "edge_excl", "=", "0", ",", "int_excl", "=", "0", ")", ":", "out1", "=", "np", ".", "empty", "(", "[", "in1", ".", "shape", "[", "0", "]", "]", ")", "mid", "=", "in1", ".", "shape", "[", "1", "]", "/", "2", "if", "(", "edge_excl", "!=", "0", ")", "|", "(", "int_excl", "!=", "0", ")", ":", "if", "edge_excl", "!=", "0", ":", "mask", "=", "np", ".", "zeros", "(", "[", "in1", ".", "shape", "[", "1", "]", ",", "in1", ".", "shape", "[", "2", "]", "]", ")", "mask", "[", "edge_excl", ":", "-", "edge_excl", ",", "edge_excl", ":", "-", "edge_excl", "]", "=", "1", "else", ":", "mask", "=", "np", ".", "ones", "(", "[", "in1", ".", "shape", "[", "1", "]", ",", "in1", ".", "shape", "[", "2", "]", "]", ")", "if", "int_excl", "!=", "0", ":", "mask", "[", "mid", "-", "int_excl", ":", "mid", "+", "int_excl", ",", "mid", "-", "int_excl", ":", "mid", "+", "int_excl", "]", "=", "0", "else", ":", "mask", "=", "np", ".", "ones", "(", "[", "in1", ".", "shape", "[", "1", "]", ",", "in1", ".", "shape", "[", "2", "]", "]", ")", "for", "i", "in", "range", "(", "in1", ".", "shape", "[", "0", "]", ")", ":", "out1", "[", "i", "]", "=", "np", ".", "median", "(", "np", ".", "abs", "(", "in1", "[", "i", ",", "mask", "==", "1", "]", ")", ")", "/", "0.6745", "return", "out1" ]
This function estimates the noise using the MAD estimator. INPUTS: in1 (no default): The array from which the noise is estimated OUTPUTS: out1 An array of per-scale noise estimates.
[ "This", "function", "estimates", "the", "noise", "using", "the", "MAD", "estimator", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_toolbox.py#L17-L48
train
ratt-ru/PyMORESANE
pymoresane/iuwt_toolbox.py
source_extraction
def source_extraction(in1, tolerance, mode="cpu", store_on_gpu=False, neg_comp=False): """ Convenience function for allocating work to cpu or gpu, depending on the selected mode. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. mode (default="cpu"):Mode of operation - either "gpu" or "cpu". OUTPUTS: Array containing the significant wavelet coefficients of extracted sources. """ if mode=="cpu": return cpu_source_extraction(in1, tolerance, neg_comp) elif mode=="gpu": return gpu_source_extraction(in1, tolerance, store_on_gpu, neg_comp)
python
def source_extraction(in1, tolerance, mode="cpu", store_on_gpu=False, neg_comp=False): """ Convenience function for allocating work to cpu or gpu, depending on the selected mode. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. mode (default="cpu"):Mode of operation - either "gpu" or "cpu". OUTPUTS: Array containing the significant wavelet coefficients of extracted sources. """ if mode=="cpu": return cpu_source_extraction(in1, tolerance, neg_comp) elif mode=="gpu": return gpu_source_extraction(in1, tolerance, store_on_gpu, neg_comp)
[ "def", "source_extraction", "(", "in1", ",", "tolerance", ",", "mode", "=", "\"cpu\"", ",", "store_on_gpu", "=", "False", ",", "neg_comp", "=", "False", ")", ":", "if", "mode", "==", "\"cpu\"", ":", "return", "cpu_source_extraction", "(", "in1", ",", "tolerance", ",", "neg_comp", ")", "elif", "mode", "==", "\"gpu\"", ":", "return", "gpu_source_extraction", "(", "in1", ",", "tolerance", ",", "store_on_gpu", ",", "neg_comp", ")" ]
Convenience function for allocating work to cpu or gpu, depending on the selected mode. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. mode (default="cpu"):Mode of operation - either "gpu" or "cpu". OUTPUTS: Array containing the significant wavelet coefficients of extracted sources.
[ "Convenience", "function", "for", "allocating", "work", "to", "cpu", "or", "gpu", "depending", "on", "the", "selected", "mode", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_toolbox.py#L77-L94
train
ratt-ru/PyMORESANE
pymoresane/iuwt_toolbox.py
cpu_source_extraction
def cpu_source_extraction(in1, tolerance, neg_comp): """ The following function determines connectivity within a given wavelet decomposition. These connected and labelled structures are thresholded to within some tolerance of the maximum coefficient at the scale. This determines whether on not an object is to be considered as significant. Significant objects are extracted and factored into a mask which is finally multiplied by the wavelet coefficients to return only wavelet coefficients belonging to significant objects across all scales. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. OUTPUTS: objects*in1 The wavelet coefficients of the significant structures. objects The mask of the significant structures. """ # The following initialises some variables for storing the labelled image and the number of labels. The per scale # maxima are also initialised here. scale_maxima = np.empty([in1.shape[0],1]) objects = np.empty_like(in1, dtype=np.int32) object_count = np.empty([in1.shape[0],1], dtype=np.int32) # The following loop uses functionality from the ndimage module to assess connectivity. The maxima are also # calculated here. for i in range(in1.shape[0]): if neg_comp: scale_maxima[i] = np.max(abs(in1[i,:,:])) else: scale_maxima[i] = np.max(in1[i,:,:]) objects[i,:,:], object_count[i] = ndimage.label(in1[i,:,:], structure=[[1,1,1],[1,1,1],[1,1,1]]) # The following removes the insignificant objects and then extracts the remaining ones. for i in range(-1,-in1.shape[0]-1,-1): if neg_comp: if i==(-1): tmp = (abs(in1[i,:,:])>=(tolerance*scale_maxima[i]))*objects[i,:,:] else: tmp = (abs(in1[i,:,:])>=(tolerance*scale_maxima[i]))*objects[i,:,:]*objects[i+1,:,:] else: if i==(-1): tmp = (in1[i,:,:]>=(tolerance*scale_maxima[i]))*objects[i,:,:] else: tmp = (in1[i,:,:]>=(tolerance*scale_maxima[i]))*objects[i,:,:]*objects[i+1,:,:] labels = np.unique(tmp[tmp>0]) for j in labels: objects[i,(objects[i,:,:]==j)] = -1 objects[i,(objects[i,:,:]>0)] = 0 objects[i,:,:] = -(objects[i,:,:]) return objects*in1, objects
python
def cpu_source_extraction(in1, tolerance, neg_comp): """ The following function determines connectivity within a given wavelet decomposition. These connected and labelled structures are thresholded to within some tolerance of the maximum coefficient at the scale. This determines whether on not an object is to be considered as significant. Significant objects are extracted and factored into a mask which is finally multiplied by the wavelet coefficients to return only wavelet coefficients belonging to significant objects across all scales. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. OUTPUTS: objects*in1 The wavelet coefficients of the significant structures. objects The mask of the significant structures. """ # The following initialises some variables for storing the labelled image and the number of labels. The per scale # maxima are also initialised here. scale_maxima = np.empty([in1.shape[0],1]) objects = np.empty_like(in1, dtype=np.int32) object_count = np.empty([in1.shape[0],1], dtype=np.int32) # The following loop uses functionality from the ndimage module to assess connectivity. The maxima are also # calculated here. for i in range(in1.shape[0]): if neg_comp: scale_maxima[i] = np.max(abs(in1[i,:,:])) else: scale_maxima[i] = np.max(in1[i,:,:]) objects[i,:,:], object_count[i] = ndimage.label(in1[i,:,:], structure=[[1,1,1],[1,1,1],[1,1,1]]) # The following removes the insignificant objects and then extracts the remaining ones. for i in range(-1,-in1.shape[0]-1,-1): if neg_comp: if i==(-1): tmp = (abs(in1[i,:,:])>=(tolerance*scale_maxima[i]))*objects[i,:,:] else: tmp = (abs(in1[i,:,:])>=(tolerance*scale_maxima[i]))*objects[i,:,:]*objects[i+1,:,:] else: if i==(-1): tmp = (in1[i,:,:]>=(tolerance*scale_maxima[i]))*objects[i,:,:] else: tmp = (in1[i,:,:]>=(tolerance*scale_maxima[i]))*objects[i,:,:]*objects[i+1,:,:] labels = np.unique(tmp[tmp>0]) for j in labels: objects[i,(objects[i,:,:]==j)] = -1 objects[i,(objects[i,:,:]>0)] = 0 objects[i,:,:] = -(objects[i,:,:]) return objects*in1, objects
[ "def", "cpu_source_extraction", "(", "in1", ",", "tolerance", ",", "neg_comp", ")", ":", "scale_maxima", "=", "np", ".", "empty", "(", "[", "in1", ".", "shape", "[", "0", "]", ",", "1", "]", ")", "objects", "=", "np", ".", "empty_like", "(", "in1", ",", "dtype", "=", "np", ".", "int32", ")", "object_count", "=", "np", ".", "empty", "(", "[", "in1", ".", "shape", "[", "0", "]", ",", "1", "]", ",", "dtype", "=", "np", ".", "int32", ")", "for", "i", "in", "range", "(", "in1", ".", "shape", "[", "0", "]", ")", ":", "if", "neg_comp", ":", "scale_maxima", "[", "i", "]", "=", "np", ".", "max", "(", "abs", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ")", ")", "else", ":", "scale_maxima", "[", "i", "]", "=", "np", ".", "max", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ")", "objects", "[", "i", ",", ":", ",", ":", "]", ",", "object_count", "[", "i", "]", "=", "ndimage", ".", "label", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ",", "structure", "=", "[", "[", "1", ",", "1", ",", "1", "]", ",", "[", "1", ",", "1", ",", "1", "]", ",", "[", "1", ",", "1", ",", "1", "]", "]", ")", "for", "i", "in", "range", "(", "-", "1", ",", "-", "in1", ".", "shape", "[", "0", "]", "-", "1", ",", "-", "1", ")", ":", "if", "neg_comp", ":", "if", "i", "==", "(", "-", "1", ")", ":", "tmp", "=", "(", "abs", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ")", ">=", "(", "tolerance", "*", "scale_maxima", "[", "i", "]", ")", ")", "*", "objects", "[", "i", ",", ":", ",", ":", "]", "else", ":", "tmp", "=", "(", "abs", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ")", ">=", "(", "tolerance", "*", "scale_maxima", "[", "i", "]", ")", ")", "*", "objects", "[", "i", ",", ":", ",", ":", "]", "*", "objects", "[", "i", "+", "1", ",", ":", ",", ":", "]", "else", ":", "if", "i", "==", "(", "-", "1", ")", ":", "tmp", "=", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ">=", "(", "tolerance", "*", "scale_maxima", "[", "i", "]", ")", ")", "*", "objects", "[", "i", ",", ":", ",", ":", "]", "else", ":", "tmp", "=", "(", "in1", "[", "i", ",", ":", ",", ":", "]", ">=", "(", "tolerance", "*", "scale_maxima", "[", "i", "]", ")", ")", "*", "objects", "[", "i", ",", ":", ",", ":", "]", "*", "objects", "[", "i", "+", "1", ",", ":", ",", ":", "]", "labels", "=", "np", ".", "unique", "(", "tmp", "[", "tmp", ">", "0", "]", ")", "for", "j", "in", "labels", ":", "objects", "[", "i", ",", "(", "objects", "[", "i", ",", ":", ",", ":", "]", "==", "j", ")", "]", "=", "-", "1", "objects", "[", "i", ",", "(", "objects", "[", "i", ",", ":", ",", ":", "]", ">", "0", ")", "]", "=", "0", "objects", "[", "i", ",", ":", ",", ":", "]", "=", "-", "(", "objects", "[", "i", ",", ":", ",", ":", "]", ")", "return", "objects", "*", "in1", ",", "objects" ]
The following function determines connectivity within a given wavelet decomposition. These connected and labelled structures are thresholded to within some tolerance of the maximum coefficient at the scale. This determines whether on not an object is to be considered as significant. Significant objects are extracted and factored into a mask which is finally multiplied by the wavelet coefficients to return only wavelet coefficients belonging to significant objects across all scales. INPUTS: in1 (no default): Array containing the wavelet decomposition. tolerance (no default): Percentage of maximum coefficient at which objects are deemed significant. OUTPUTS: objects*in1 The wavelet coefficients of the significant structures. objects The mask of the significant structures.
[ "The", "following", "function", "determines", "connectivity", "within", "a", "given", "wavelet", "decomposition", ".", "These", "connected", "and", "labelled", "structures", "are", "thresholded", "to", "within", "some", "tolerance", "of", "the", "maximum", "coefficient", "at", "the", "scale", ".", "This", "determines", "whether", "on", "not", "an", "object", "is", "to", "be", "considered", "as", "significant", ".", "Significant", "objects", "are", "extracted", "and", "factored", "into", "a", "mask", "which", "is", "finally", "multiplied", "by", "the", "wavelet", "coefficients", "to", "return", "only", "wavelet", "coefficients", "belonging", "to", "significant", "objects", "across", "all", "scales", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_toolbox.py#L97-L154
train
ratt-ru/PyMORESANE
pymoresane/iuwt_toolbox.py
snr_ratio
def snr_ratio(in1, in2): """ The following function simply calculates the signal to noise ratio between two signals. INPUTS: in1 (no default): Array containing values for signal 1. in2 (no default): Array containing values for signal 2. OUTPUTS: out1 The ratio of the signal to noise ratios of two signals. """ out1 = 20*(np.log10(np.linalg.norm(in1)/np.linalg.norm(in1-in2))) return out1
python
def snr_ratio(in1, in2): """ The following function simply calculates the signal to noise ratio between two signals. INPUTS: in1 (no default): Array containing values for signal 1. in2 (no default): Array containing values for signal 2. OUTPUTS: out1 The ratio of the signal to noise ratios of two signals. """ out1 = 20*(np.log10(np.linalg.norm(in1)/np.linalg.norm(in1-in2))) return out1
[ "def", "snr_ratio", "(", "in1", ",", "in2", ")", ":", "out1", "=", "20", "*", "(", "np", ".", "log10", "(", "np", ".", "linalg", ".", "norm", "(", "in1", ")", "/", "np", ".", "linalg", ".", "norm", "(", "in1", "-", "in2", ")", ")", ")", "return", "out1" ]
The following function simply calculates the signal to noise ratio between two signals. INPUTS: in1 (no default): Array containing values for signal 1. in2 (no default): Array containing values for signal 2. OUTPUTS: out1 The ratio of the signal to noise ratios of two signals.
[ "The", "following", "function", "simply", "calculates", "the", "signal", "to", "noise", "ratio", "between", "two", "signals", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_toolbox.py#L292-L306
train
praekeltfoundation/seaworthy
seaworthy/containers/rabbitmq.py
RabbitMQContainer.wait_for_start
def wait_for_start(self): """ Wait for the RabbitMQ process to be come up. """ er = self.exec_rabbitmqctl( 'wait', ['--pid', '1', '--timeout', str(int(self.wait_timeout))]) output_lines(er, error_exc=TimeoutError)
python
def wait_for_start(self): """ Wait for the RabbitMQ process to be come up. """ er = self.exec_rabbitmqctl( 'wait', ['--pid', '1', '--timeout', str(int(self.wait_timeout))]) output_lines(er, error_exc=TimeoutError)
[ "def", "wait_for_start", "(", "self", ")", ":", "er", "=", "self", ".", "exec_rabbitmqctl", "(", "'wait'", ",", "[", "'--pid'", ",", "'1'", ",", "'--timeout'", ",", "str", "(", "int", "(", "self", ".", "wait_timeout", ")", ")", "]", ")", "output_lines", "(", "er", ",", "error_exc", "=", "TimeoutError", ")" ]
Wait for the RabbitMQ process to be come up.
[ "Wait", "for", "the", "RabbitMQ", "process", "to", "be", "come", "up", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/rabbitmq.py#L56-L62
train
praekeltfoundation/seaworthy
seaworthy/containers/rabbitmq.py
RabbitMQContainer.exec_rabbitmqctl
def exec_rabbitmqctl(self, command, args=[], rabbitmqctl_opts=['-q']): """ Execute a ``rabbitmqctl`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output """ cmd = ['rabbitmqctl'] + rabbitmqctl_opts + [command] + args return self.inner().exec_run(cmd)
python
def exec_rabbitmqctl(self, command, args=[], rabbitmqctl_opts=['-q']): """ Execute a ``rabbitmqctl`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output """ cmd = ['rabbitmqctl'] + rabbitmqctl_opts + [command] + args return self.inner().exec_run(cmd)
[ "def", "exec_rabbitmqctl", "(", "self", ",", "command", ",", "args", "=", "[", "]", ",", "rabbitmqctl_opts", "=", "[", "'-q'", "]", ")", ":", "cmd", "=", "[", "'rabbitmqctl'", "]", "+", "rabbitmqctl_opts", "+", "[", "command", "]", "+", "args", "return", "self", ".", "inner", "(", ")", ".", "exec_run", "(", "cmd", ")" ]
Execute a ``rabbitmqctl`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output
[ "Execute", "a", "rabbitmqctl", "command", "inside", "a", "running", "container", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/rabbitmq.py#L87-L98
train
praekeltfoundation/seaworthy
seaworthy/containers/rabbitmq.py
RabbitMQContainer.exec_rabbitmqctl_list
def exec_rabbitmqctl_list(self, resources, args=[], rabbitmq_opts=['-q', '--no-table-headers']): """ Execute a ``rabbitmqctl`` command to list the given resources. :param resources: the resources to list, e.g. ``'vhosts'`` :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output """ command = 'list_{}'.format(resources) return self.exec_rabbitmqctl(command, args, rabbitmq_opts)
python
def exec_rabbitmqctl_list(self, resources, args=[], rabbitmq_opts=['-q', '--no-table-headers']): """ Execute a ``rabbitmqctl`` command to list the given resources. :param resources: the resources to list, e.g. ``'vhosts'`` :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output """ command = 'list_{}'.format(resources) return self.exec_rabbitmqctl(command, args, rabbitmq_opts)
[ "def", "exec_rabbitmqctl_list", "(", "self", ",", "resources", ",", "args", "=", "[", "]", ",", "rabbitmq_opts", "=", "[", "'-q'", ",", "'--no-table-headers'", "]", ")", ":", "command", "=", "'list_{}'", ".", "format", "(", "resources", ")", "return", "self", ".", "exec_rabbitmqctl", "(", "command", ",", "args", ",", "rabbitmq_opts", ")" ]
Execute a ``rabbitmqctl`` command to list the given resources. :param resources: the resources to list, e.g. ``'vhosts'`` :param args: a list of args for the command :param rabbitmqctl_opts: a list of extra options to pass to ``rabbitmqctl`` :returns: a tuple of the command exit code and output
[ "Execute", "a", "rabbitmqctl", "command", "to", "list", "the", "given", "resources", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/rabbitmq.py#L100-L112
train
praekeltfoundation/seaworthy
seaworthy/containers/rabbitmq.py
RabbitMQContainer.list_users
def list_users(self): """ Run the ``list_users`` command and return a list of tuples describing the users. :return: A list of 2-element tuples. The first element is the username, the second a list of tags for the user. """ lines = output_lines(self.exec_rabbitmqctl_list('users')) return [_parse_rabbitmq_user(line) for line in lines]
python
def list_users(self): """ Run the ``list_users`` command and return a list of tuples describing the users. :return: A list of 2-element tuples. The first element is the username, the second a list of tags for the user. """ lines = output_lines(self.exec_rabbitmqctl_list('users')) return [_parse_rabbitmq_user(line) for line in lines]
[ "def", "list_users", "(", "self", ")", ":", "lines", "=", "output_lines", "(", "self", ".", "exec_rabbitmqctl_list", "(", "'users'", ")", ")", "return", "[", "_parse_rabbitmq_user", "(", "line", ")", "for", "line", "in", "lines", "]" ]
Run the ``list_users`` command and return a list of tuples describing the users. :return: A list of 2-element tuples. The first element is the username, the second a list of tags for the user.
[ "Run", "the", "list_users", "command", "and", "return", "a", "list", "of", "tuples", "describing", "the", "users", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/rabbitmq.py#L133-L143
train
praekeltfoundation/seaworthy
seaworthy/containers/rabbitmq.py
RabbitMQContainer.broker_url
def broker_url(self): """ Returns a "broker URL" for use with Celery. """ return 'amqp://{}:{}@{}/{}'.format( self.user, self.password, self.name, self.vhost)
python
def broker_url(self): """ Returns a "broker URL" for use with Celery. """ return 'amqp://{}:{}@{}/{}'.format( self.user, self.password, self.name, self.vhost)
[ "def", "broker_url", "(", "self", ")", ":", "return", "'amqp://{}:{}@{}/{}'", ".", "format", "(", "self", ".", "user", ",", "self", ".", "password", ",", "self", ".", "name", ",", "self", ".", "vhost", ")" ]
Returns a "broker URL" for use with Celery.
[ "Returns", "a", "broker", "URL", "for", "use", "with", "Celery", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/rabbitmq.py#L145-L148
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.exec_pg_success
def exec_pg_success(self, cmd): """ Execute a command inside a running container as the postgres user, asserting success. """ result = self.inner().exec_run(cmd, user='postgres') assert result.exit_code == 0, result.output.decode('utf-8') return result
python
def exec_pg_success(self, cmd): """ Execute a command inside a running container as the postgres user, asserting success. """ result = self.inner().exec_run(cmd, user='postgres') assert result.exit_code == 0, result.output.decode('utf-8') return result
[ "def", "exec_pg_success", "(", "self", ",", "cmd", ")", ":", "result", "=", "self", ".", "inner", "(", ")", ".", "exec_run", "(", "cmd", ",", "user", "=", "'postgres'", ")", "assert", "result", ".", "exit_code", "==", "0", ",", "result", ".", "output", ".", "decode", "(", "'utf-8'", ")", "return", "result" ]
Execute a command inside a running container as the postgres user, asserting success.
[ "Execute", "a", "command", "inside", "a", "running", "container", "as", "the", "postgres", "user", "asserting", "success", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L63-L70
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.clean
def clean(self): """ Remove all data by dropping and recreating the configured database. .. note:: Only the configured database is removed. Any other databases remain untouched. """ self.exec_pg_success(['dropdb', '-U', self.user, self.database]) self.exec_pg_success(['createdb', '-U', self.user, self.database])
python
def clean(self): """ Remove all data by dropping and recreating the configured database. .. note:: Only the configured database is removed. Any other databases remain untouched. """ self.exec_pg_success(['dropdb', '-U', self.user, self.database]) self.exec_pg_success(['createdb', '-U', self.user, self.database])
[ "def", "clean", "(", "self", ")", ":", "self", ".", "exec_pg_success", "(", "[", "'dropdb'", ",", "'-U'", ",", "self", ".", "user", ",", "self", ".", "database", "]", ")", "self", ".", "exec_pg_success", "(", "[", "'createdb'", ",", "'-U'", ",", "self", ".", "user", ",", "self", ".", "database", "]", ")" ]
Remove all data by dropping and recreating the configured database. .. note:: Only the configured database is removed. Any other databases remain untouched.
[ "Remove", "all", "data", "by", "dropping", "and", "recreating", "the", "configured", "database", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L72-L82
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.exec_psql
def exec_psql(self, command, psql_opts=['-qtA']): """ Execute a ``psql`` command inside a running container. By default the container's database is connected to. :param command: the command to run (passed to ``-c``) :param psql_opts: a list of extra options to pass to ``psql`` :returns: a tuple of the command exit code and output """ cmd = ['psql'] + psql_opts + [ '--dbname', self.database, '-U', self.user, '-c', command, ] return self.inner().exec_run(cmd, user='postgres')
python
def exec_psql(self, command, psql_opts=['-qtA']): """ Execute a ``psql`` command inside a running container. By default the container's database is connected to. :param command: the command to run (passed to ``-c``) :param psql_opts: a list of extra options to pass to ``psql`` :returns: a tuple of the command exit code and output """ cmd = ['psql'] + psql_opts + [ '--dbname', self.database, '-U', self.user, '-c', command, ] return self.inner().exec_run(cmd, user='postgres')
[ "def", "exec_psql", "(", "self", ",", "command", ",", "psql_opts", "=", "[", "'-qtA'", "]", ")", ":", "cmd", "=", "[", "'psql'", "]", "+", "psql_opts", "+", "[", "'--dbname'", ",", "self", ".", "database", ",", "'-U'", ",", "self", ".", "user", ",", "'-c'", ",", "command", ",", "]", "return", "self", ".", "inner", "(", ")", ".", "exec_run", "(", "cmd", ",", "user", "=", "'postgres'", ")" ]
Execute a ``psql`` command inside a running container. By default the container's database is connected to. :param command: the command to run (passed to ``-c``) :param psql_opts: a list of extra options to pass to ``psql`` :returns: a tuple of the command exit code and output
[ "Execute", "a", "psql", "command", "inside", "a", "running", "container", ".", "By", "default", "the", "container", "s", "database", "is", "connected", "to", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L84-L98
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.list_databases
def list_databases(self): """ Runs the ``\\list`` command and returns a list of column values with information about all databases. """ lines = output_lines(self.exec_psql('\\list')) return [line.split('|') for line in lines]
python
def list_databases(self): """ Runs the ``\\list`` command and returns a list of column values with information about all databases. """ lines = output_lines(self.exec_psql('\\list')) return [line.split('|') for line in lines]
[ "def", "list_databases", "(", "self", ")", ":", "lines", "=", "output_lines", "(", "self", ".", "exec_psql", "(", "'\\\\list'", ")", ")", "return", "[", "line", ".", "split", "(", "'|'", ")", "for", "line", "in", "lines", "]" ]
Runs the ``\\list`` command and returns a list of column values with information about all databases.
[ "Runs", "the", "\\\\", "list", "command", "and", "returns", "a", "list", "of", "column", "values", "with", "information", "about", "all", "databases", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L100-L106
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.list_tables
def list_tables(self): """ Runs the ``\\dt`` command and returns a list of column values with information about all tables in the database. """ lines = output_lines(self.exec_psql('\\dt')) return [line.split('|') for line in lines]
python
def list_tables(self): """ Runs the ``\\dt`` command and returns a list of column values with information about all tables in the database. """ lines = output_lines(self.exec_psql('\\dt')) return [line.split('|') for line in lines]
[ "def", "list_tables", "(", "self", ")", ":", "lines", "=", "output_lines", "(", "self", ".", "exec_psql", "(", "'\\\\dt'", ")", ")", "return", "[", "line", ".", "split", "(", "'|'", ")", "for", "line", "in", "lines", "]" ]
Runs the ``\\dt`` command and returns a list of column values with information about all tables in the database.
[ "Runs", "the", "\\\\", "dt", "command", "and", "returns", "a", "list", "of", "column", "values", "with", "information", "about", "all", "tables", "in", "the", "database", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L108-L114
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.list_users
def list_users(self): """ Runs the ``\\du`` command and returns a list of column values with information about all user roles. """ lines = output_lines(self.exec_psql('\\du')) return [line.split('|') for line in lines]
python
def list_users(self): """ Runs the ``\\du`` command and returns a list of column values with information about all user roles. """ lines = output_lines(self.exec_psql('\\du')) return [line.split('|') for line in lines]
[ "def", "list_users", "(", "self", ")", ":", "lines", "=", "output_lines", "(", "self", ".", "exec_psql", "(", "'\\\\du'", ")", ")", "return", "[", "line", ".", "split", "(", "'|'", ")", "for", "line", "in", "lines", "]" ]
Runs the ``\\du`` command and returns a list of column values with information about all user roles.
[ "Runs", "the", "\\\\", "du", "command", "and", "returns", "a", "list", "of", "column", "values", "with", "information", "about", "all", "user", "roles", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L116-L122
train
praekeltfoundation/seaworthy
seaworthy/containers/postgresql.py
PostgreSQLContainer.database_url
def database_url(self): """ Returns a "database URL" for use with DJ-Database-URL and similar libraries. """ return 'postgres://{}:{}@{}/{}'.format( self.user, self.password, self.name, self.database)
python
def database_url(self): """ Returns a "database URL" for use with DJ-Database-URL and similar libraries. """ return 'postgres://{}:{}@{}/{}'.format( self.user, self.password, self.name, self.database)
[ "def", "database_url", "(", "self", ")", ":", "return", "'postgres://{}:{}@{}/{}'", ".", "format", "(", "self", ".", "user", ",", "self", ".", "password", ",", "self", ".", "name", ",", "self", ".", "database", ")" ]
Returns a "database URL" for use with DJ-Database-URL and similar libraries.
[ "Returns", "a", "database", "URL", "for", "use", "with", "DJ", "-", "Database", "-", "URL", "and", "similar", "libraries", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/postgresql.py#L124-L130
train
ionelmc/python-matrix
src/matrix/__init__.py
from_config
def from_config(config): """ Generate a matrix from a configuration dictionary. """ matrix = {} variables = config.keys() for entries in product(*config.values()): combination = dict(zip(variables, entries)) include = True for value in combination.values(): for reducer in value.reducers: if reducer.pattern == '-': match = not combination[reducer.variable].value else: match = fnmatch(combination[reducer.variable].value, reducer.pattern) if match if reducer.is_exclude else not match: include = False if include: key = '-'.join(entry.alias for entry in entries if entry.alias) data = dict( zip(variables, (entry.value for entry in entries)) ) if key in matrix and data != matrix[key]: raise DuplicateEnvironment(key, data, matrix[key]) matrix[key] = data return matrix
python
def from_config(config): """ Generate a matrix from a configuration dictionary. """ matrix = {} variables = config.keys() for entries in product(*config.values()): combination = dict(zip(variables, entries)) include = True for value in combination.values(): for reducer in value.reducers: if reducer.pattern == '-': match = not combination[reducer.variable].value else: match = fnmatch(combination[reducer.variable].value, reducer.pattern) if match if reducer.is_exclude else not match: include = False if include: key = '-'.join(entry.alias for entry in entries if entry.alias) data = dict( zip(variables, (entry.value for entry in entries)) ) if key in matrix and data != matrix[key]: raise DuplicateEnvironment(key, data, matrix[key]) matrix[key] = data return matrix
[ "def", "from_config", "(", "config", ")", ":", "matrix", "=", "{", "}", "variables", "=", "config", ".", "keys", "(", ")", "for", "entries", "in", "product", "(", "*", "config", ".", "values", "(", ")", ")", ":", "combination", "=", "dict", "(", "zip", "(", "variables", ",", "entries", ")", ")", "include", "=", "True", "for", "value", "in", "combination", ".", "values", "(", ")", ":", "for", "reducer", "in", "value", ".", "reducers", ":", "if", "reducer", ".", "pattern", "==", "'-'", ":", "match", "=", "not", "combination", "[", "reducer", ".", "variable", "]", ".", "value", "else", ":", "match", "=", "fnmatch", "(", "combination", "[", "reducer", ".", "variable", "]", ".", "value", ",", "reducer", ".", "pattern", ")", "if", "match", "if", "reducer", ".", "is_exclude", "else", "not", "match", ":", "include", "=", "False", "if", "include", ":", "key", "=", "'-'", ".", "join", "(", "entry", ".", "alias", "for", "entry", "in", "entries", "if", "entry", ".", "alias", ")", "data", "=", "dict", "(", "zip", "(", "variables", ",", "(", "entry", ".", "value", "for", "entry", "in", "entries", ")", ")", ")", "if", "key", "in", "matrix", "and", "data", "!=", "matrix", "[", "key", "]", ":", "raise", "DuplicateEnvironment", "(", "key", ",", "data", ",", "matrix", "[", "key", "]", ")", "matrix", "[", "key", "]", "=", "data", "return", "matrix" ]
Generate a matrix from a configuration dictionary.
[ "Generate", "a", "matrix", "from", "a", "configuration", "dictionary", "." ]
e1a63879a6c94c37c3883386f1d86eb7c2179a5b
https://github.com/ionelmc/python-matrix/blob/e1a63879a6c94c37c3883386f1d86eb7c2179a5b/src/matrix/__init__.py#L131-L156
train
tuomas2/automate
src/automate/worker.py
StatusWorkerThread.flush
def flush(self): """ This only needs to be called manually from unit tests """ self.logger.debug('Flush joining') self.queue.join() self.logger.debug('Flush joining ready')
python
def flush(self): """ This only needs to be called manually from unit tests """ self.logger.debug('Flush joining') self.queue.join() self.logger.debug('Flush joining ready')
[ "def", "flush", "(", "self", ")", ":", "self", ".", "logger", ".", "debug", "(", "'Flush joining'", ")", "self", ".", "queue", ".", "join", "(", ")", "self", ".", "logger", ".", "debug", "(", "'Flush joining ready'", ")" ]
This only needs to be called manually from unit tests
[ "This", "only", "needs", "to", "be", "called", "manually", "from", "unit", "tests" ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/worker.py#L112-L119
train
praekeltfoundation/seaworthy
seaworthy/utils.py
output_lines
def output_lines(output, encoding='utf-8', error_exc=None): """ Convert bytestring container output or the result of a container exec command into a sequence of unicode lines. :param output: Container output bytes or an :class:`docker.models.containers.ExecResult` instance. :param encoding: The encoding to use when converting bytes to unicode (default ``utf-8``). :param error_exc: Optional exception to raise if ``output`` is an ``ExecResult`` with a nonzero exit code. :returns: list[str] """ if isinstance(output, ExecResult): exit_code, output = output if exit_code != 0 and error_exc is not None: raise error_exc(output.decode(encoding)) return output.decode(encoding).splitlines()
python
def output_lines(output, encoding='utf-8', error_exc=None): """ Convert bytestring container output or the result of a container exec command into a sequence of unicode lines. :param output: Container output bytes or an :class:`docker.models.containers.ExecResult` instance. :param encoding: The encoding to use when converting bytes to unicode (default ``utf-8``). :param error_exc: Optional exception to raise if ``output`` is an ``ExecResult`` with a nonzero exit code. :returns: list[str] """ if isinstance(output, ExecResult): exit_code, output = output if exit_code != 0 and error_exc is not None: raise error_exc(output.decode(encoding)) return output.decode(encoding).splitlines()
[ "def", "output_lines", "(", "output", ",", "encoding", "=", "'utf-8'", ",", "error_exc", "=", "None", ")", ":", "if", "isinstance", "(", "output", ",", "ExecResult", ")", ":", "exit_code", ",", "output", "=", "output", "if", "exit_code", "!=", "0", "and", "error_exc", "is", "not", "None", ":", "raise", "error_exc", "(", "output", ".", "decode", "(", "encoding", ")", ")", "return", "output", ".", "decode", "(", "encoding", ")", ".", "splitlines", "(", ")" ]
Convert bytestring container output or the result of a container exec command into a sequence of unicode lines. :param output: Container output bytes or an :class:`docker.models.containers.ExecResult` instance. :param encoding: The encoding to use when converting bytes to unicode (default ``utf-8``). :param error_exc: Optional exception to raise if ``output`` is an ``ExecResult`` with a nonzero exit code. :returns: list[str]
[ "Convert", "bytestring", "container", "output", "or", "the", "result", "of", "a", "container", "exec", "command", "into", "a", "sequence", "of", "unicode", "lines", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/utils.py#L4-L27
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.get_file
def get_file(self, fid): """Get file from WeedFS. Returns file content. May be problematic for large files as content is stored in memory. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Content of the file with provided fid or None if file doesn't exist on the server .. versionadded:: 0.3.1 """ url = self.get_file_url(fid) return self.conn.get_raw_data(url)
python
def get_file(self, fid): """Get file from WeedFS. Returns file content. May be problematic for large files as content is stored in memory. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Content of the file with provided fid or None if file doesn't exist on the server .. versionadded:: 0.3.1 """ url = self.get_file_url(fid) return self.conn.get_raw_data(url)
[ "def", "get_file", "(", "self", ",", "fid", ")", ":", "url", "=", "self", ".", "get_file_url", "(", "fid", ")", "return", "self", ".", "conn", ".", "get_raw_data", "(", "url", ")" ]
Get file from WeedFS. Returns file content. May be problematic for large files as content is stored in memory. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Content of the file with provided fid or None if file doesn't exist on the server .. versionadded:: 0.3.1
[ "Get", "file", "from", "WeedFS", "." ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L50-L66
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.get_file_url
def get_file_url(self, fid, public=None): """ Get url for the file :param string fid: File ID :param boolean public: public or internal url :rtype: string """ try: volume_id, rest = fid.strip().split(",") except ValueError: raise BadFidFormat( "fid must be in format: <volume_id>,<file_name_hash>") file_location = self.get_file_location(volume_id) if public is None: public = self.use_public_url volume_url = file_location.public_url if public else file_location.url url = "http://{volume_url}/{fid}".format( volume_url=volume_url, fid=fid) return url
python
def get_file_url(self, fid, public=None): """ Get url for the file :param string fid: File ID :param boolean public: public or internal url :rtype: string """ try: volume_id, rest = fid.strip().split(",") except ValueError: raise BadFidFormat( "fid must be in format: <volume_id>,<file_name_hash>") file_location = self.get_file_location(volume_id) if public is None: public = self.use_public_url volume_url = file_location.public_url if public else file_location.url url = "http://{volume_url}/{fid}".format( volume_url=volume_url, fid=fid) return url
[ "def", "get_file_url", "(", "self", ",", "fid", ",", "public", "=", "None", ")", ":", "try", ":", "volume_id", ",", "rest", "=", "fid", ".", "strip", "(", ")", ".", "split", "(", "\",\"", ")", "except", "ValueError", ":", "raise", "BadFidFormat", "(", "\"fid must be in format: <volume_id>,<file_name_hash>\"", ")", "file_location", "=", "self", ".", "get_file_location", "(", "volume_id", ")", "if", "public", "is", "None", ":", "public", "=", "self", ".", "use_public_url", "volume_url", "=", "file_location", ".", "public_url", "if", "public", "else", "file_location", ".", "url", "url", "=", "\"http://{volume_url}/{fid}\"", ".", "format", "(", "volume_url", "=", "volume_url", ",", "fid", "=", "fid", ")", "return", "url" ]
Get url for the file :param string fid: File ID :param boolean public: public or internal url :rtype: string
[ "Get", "url", "for", "the", "file" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L68-L87
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.get_file_location
def get_file_location(self, volume_id): """ Get location for the file, WeedFS volume is choosed randomly :param integer volume_id: volume_id :rtype: namedtuple `FileLocation` `{"public_url":"", "url":""}` """ url = ("http://{master_addr}:{master_port}/" "dir/lookup?volumeId={volume_id}").format( master_addr=self.master_addr, master_port=self.master_port, volume_id=volume_id) data = json.loads(self.conn.get_data(url)) _file_location = random.choice(data['locations']) FileLocation = namedtuple('FileLocation', "public_url url") return FileLocation(_file_location['publicUrl'], _file_location['url'])
python
def get_file_location(self, volume_id): """ Get location for the file, WeedFS volume is choosed randomly :param integer volume_id: volume_id :rtype: namedtuple `FileLocation` `{"public_url":"", "url":""}` """ url = ("http://{master_addr}:{master_port}/" "dir/lookup?volumeId={volume_id}").format( master_addr=self.master_addr, master_port=self.master_port, volume_id=volume_id) data = json.loads(self.conn.get_data(url)) _file_location = random.choice(data['locations']) FileLocation = namedtuple('FileLocation', "public_url url") return FileLocation(_file_location['publicUrl'], _file_location['url'])
[ "def", "get_file_location", "(", "self", ",", "volume_id", ")", ":", "url", "=", "(", "\"http://{master_addr}:{master_port}/\"", "\"dir/lookup?volumeId={volume_id}\"", ")", ".", "format", "(", "master_addr", "=", "self", ".", "master_addr", ",", "master_port", "=", "self", ".", "master_port", ",", "volume_id", "=", "volume_id", ")", "data", "=", "json", ".", "loads", "(", "self", ".", "conn", ".", "get_data", "(", "url", ")", ")", "_file_location", "=", "random", ".", "choice", "(", "data", "[", "'locations'", "]", ")", "FileLocation", "=", "namedtuple", "(", "'FileLocation'", ",", "\"public_url url\"", ")", "return", "FileLocation", "(", "_file_location", "[", "'publicUrl'", "]", ",", "_file_location", "[", "'url'", "]", ")" ]
Get location for the file, WeedFS volume is choosed randomly :param integer volume_id: volume_id :rtype: namedtuple `FileLocation` `{"public_url":"", "url":""}`
[ "Get", "location", "for", "the", "file", "WeedFS", "volume", "is", "choosed", "randomly" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L89-L105
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.get_file_size
def get_file_size(self, fid): """ Gets size of uploaded file Or None if file doesn't exist. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Int or None """ url = self.get_file_url(fid) res = self.conn.head(url) if res is not None: size = res.headers.get("content-length", None) if size is not None: return int(size) return None
python
def get_file_size(self, fid): """ Gets size of uploaded file Or None if file doesn't exist. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Int or None """ url = self.get_file_url(fid) res = self.conn.head(url) if res is not None: size = res.headers.get("content-length", None) if size is not None: return int(size) return None
[ "def", "get_file_size", "(", "self", ",", "fid", ")", ":", "url", "=", "self", ".", "get_file_url", "(", "fid", ")", "res", "=", "self", ".", "conn", ".", "head", "(", "url", ")", "if", "res", "is", "not", "None", ":", "size", "=", "res", ".", "headers", ".", "get", "(", "\"content-length\"", ",", "None", ")", "if", "size", "is", "not", "None", ":", "return", "int", "(", "size", ")", "return", "None" ]
Gets size of uploaded file Or None if file doesn't exist. Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: Int or None
[ "Gets", "size", "of", "uploaded", "file", "Or", "None", "if", "file", "doesn", "t", "exist", "." ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L107-L124
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.file_exists
def file_exists(self, fid): """Checks if file with provided fid exists Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: True if file exists. False if not. """ res = self.get_file_size(fid) if res is not None: return True return False
python
def file_exists(self, fid): """Checks if file with provided fid exists Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: True if file exists. False if not. """ res = self.get_file_size(fid) if res is not None: return True return False
[ "def", "file_exists", "(", "self", ",", "fid", ")", ":", "res", "=", "self", ".", "get_file_size", "(", "fid", ")", "if", "res", "is", "not", "None", ":", "return", "True", "return", "False" ]
Checks if file with provided fid exists Args: **fid**: File identifier <volume_id>,<file_name_hash> Returns: True if file exists. False if not.
[ "Checks", "if", "file", "with", "provided", "fid", "exists" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L126-L138
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.delete_file
def delete_file(self, fid): """ Delete file from WeedFS :param string fid: File ID """ url = self.get_file_url(fid) return self.conn.delete_data(url)
python
def delete_file(self, fid): """ Delete file from WeedFS :param string fid: File ID """ url = self.get_file_url(fid) return self.conn.delete_data(url)
[ "def", "delete_file", "(", "self", ",", "fid", ")", ":", "url", "=", "self", ".", "get_file_url", "(", "fid", ")", "return", "self", ".", "conn", ".", "delete_data", "(", "url", ")" ]
Delete file from WeedFS :param string fid: File ID
[ "Delete", "file", "from", "WeedFS" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L140-L147
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.upload_file
def upload_file(self, path=None, stream=None, name=None, **kwargs): """ Uploads file to WeedFS I takes either path or stream and name and upload it to WeedFS server. Returns fid of the uploaded file. :param string path: :param string stream: :param string name: :rtype: string or None """ params = "&".join(["%s=%s" % (k, v) for k, v in kwargs.items()]) url = "http://{master_addr}:{master_port}/dir/assign{params}".format( master_addr=self.master_addr, master_port=self.master_port, params="?" + params if params else '' ) data = json.loads(self.conn.get_data(url)) if data.get("error") is not None: return None post_url = "http://{url}/{fid}".format( url=data['publicUrl' if self.use_public_url else 'url'], fid=data['fid'] ) if path is not None: filename = os.path.basename(path) with open(path, "rb") as file_stream: res = self.conn.post_file(post_url, filename, file_stream) # we have file like object and filename elif stream is not None and name is not None: res = self.conn.post_file(post_url, name, stream) else: raise ValueError( "If `path` is None then *both* `stream` and `name` must not" " be None ") response_data = json.loads(res) if "size" in response_data: return data.get('fid') return None
python
def upload_file(self, path=None, stream=None, name=None, **kwargs): """ Uploads file to WeedFS I takes either path or stream and name and upload it to WeedFS server. Returns fid of the uploaded file. :param string path: :param string stream: :param string name: :rtype: string or None """ params = "&".join(["%s=%s" % (k, v) for k, v in kwargs.items()]) url = "http://{master_addr}:{master_port}/dir/assign{params}".format( master_addr=self.master_addr, master_port=self.master_port, params="?" + params if params else '' ) data = json.loads(self.conn.get_data(url)) if data.get("error") is not None: return None post_url = "http://{url}/{fid}".format( url=data['publicUrl' if self.use_public_url else 'url'], fid=data['fid'] ) if path is not None: filename = os.path.basename(path) with open(path, "rb") as file_stream: res = self.conn.post_file(post_url, filename, file_stream) # we have file like object and filename elif stream is not None and name is not None: res = self.conn.post_file(post_url, name, stream) else: raise ValueError( "If `path` is None then *both* `stream` and `name` must not" " be None ") response_data = json.loads(res) if "size" in response_data: return data.get('fid') return None
[ "def", "upload_file", "(", "self", ",", "path", "=", "None", ",", "stream", "=", "None", ",", "name", "=", "None", ",", "**", "kwargs", ")", ":", "params", "=", "\"&\"", ".", "join", "(", "[", "\"%s=%s\"", "%", "(", "k", ",", "v", ")", "for", "k", ",", "v", "in", "kwargs", ".", "items", "(", ")", "]", ")", "url", "=", "\"http://{master_addr}:{master_port}/dir/assign{params}\"", ".", "format", "(", "master_addr", "=", "self", ".", "master_addr", ",", "master_port", "=", "self", ".", "master_port", ",", "params", "=", "\"?\"", "+", "params", "if", "params", "else", "''", ")", "data", "=", "json", ".", "loads", "(", "self", ".", "conn", ".", "get_data", "(", "url", ")", ")", "if", "data", ".", "get", "(", "\"error\"", ")", "is", "not", "None", ":", "return", "None", "post_url", "=", "\"http://{url}/{fid}\"", ".", "format", "(", "url", "=", "data", "[", "'publicUrl'", "if", "self", ".", "use_public_url", "else", "'url'", "]", ",", "fid", "=", "data", "[", "'fid'", "]", ")", "if", "path", "is", "not", "None", ":", "filename", "=", "os", ".", "path", ".", "basename", "(", "path", ")", "with", "open", "(", "path", ",", "\"rb\"", ")", "as", "file_stream", ":", "res", "=", "self", ".", "conn", ".", "post_file", "(", "post_url", ",", "filename", ",", "file_stream", ")", "elif", "stream", "is", "not", "None", "and", "name", "is", "not", "None", ":", "res", "=", "self", ".", "conn", ".", "post_file", "(", "post_url", ",", "name", ",", "stream", ")", "else", ":", "raise", "ValueError", "(", "\"If `path` is None then *both* `stream` and `name` must not\"", "\" be None \"", ")", "response_data", "=", "json", ".", "loads", "(", "res", ")", "if", "\"size\"", "in", "response_data", ":", "return", "data", ".", "get", "(", "'fid'", ")", "return", "None" ]
Uploads file to WeedFS I takes either path or stream and name and upload it to WeedFS server. Returns fid of the uploaded file. :param string path: :param string stream: :param string name: :rtype: string or None
[ "Uploads", "file", "to", "WeedFS" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L149-L192
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.vacuum
def vacuum(self, threshold=0.3): ''' Force garbage collection :param float threshold (optional): The threshold is optional, and will not change the default threshold. :rtype: boolean ''' url = ("http://{master_addr}:{master_port}/" "vol/vacuum?garbageThreshold={threshold}").format( master_addr=self.master_addr, master_port=self.master_port, threshold=threshold) res = self.conn.get_data(url) if res is not None: return True return False
python
def vacuum(self, threshold=0.3): ''' Force garbage collection :param float threshold (optional): The threshold is optional, and will not change the default threshold. :rtype: boolean ''' url = ("http://{master_addr}:{master_port}/" "vol/vacuum?garbageThreshold={threshold}").format( master_addr=self.master_addr, master_port=self.master_port, threshold=threshold) res = self.conn.get_data(url) if res is not None: return True return False
[ "def", "vacuum", "(", "self", ",", "threshold", "=", "0.3", ")", ":", "url", "=", "(", "\"http://{master_addr}:{master_port}/\"", "\"vol/vacuum?garbageThreshold={threshold}\"", ")", ".", "format", "(", "master_addr", "=", "self", ".", "master_addr", ",", "master_port", "=", "self", ".", "master_port", ",", "threshold", "=", "threshold", ")", "res", "=", "self", ".", "conn", ".", "get_data", "(", "url", ")", "if", "res", "is", "not", "None", ":", "return", "True", "return", "False" ]
Force garbage collection :param float threshold (optional): The threshold is optional, and will not change the default threshold. :rtype: boolean
[ "Force", "garbage", "collection" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L194-L211
train
utek/pyseaweed
pyseaweed/weed.py
WeedFS.version
def version(self): ''' Returns Weed-FS master version :rtype: string ''' url = "http://{master_addr}:{master_port}/dir/status".format( master_addr=self.master_addr, master_port=self.master_port) data = self.conn.get_data(url) response_data = json.loads(data) return response_data.get("Version")
python
def version(self): ''' Returns Weed-FS master version :rtype: string ''' url = "http://{master_addr}:{master_port}/dir/status".format( master_addr=self.master_addr, master_port=self.master_port) data = self.conn.get_data(url) response_data = json.loads(data) return response_data.get("Version")
[ "def", "version", "(", "self", ")", ":", "url", "=", "\"http://{master_addr}:{master_port}/dir/status\"", ".", "format", "(", "master_addr", "=", "self", ".", "master_addr", ",", "master_port", "=", "self", ".", "master_port", ")", "data", "=", "self", ".", "conn", ".", "get_data", "(", "url", ")", "response_data", "=", "json", ".", "loads", "(", "data", ")", "return", "response_data", ".", "get", "(", "\"Version\"", ")" ]
Returns Weed-FS master version :rtype: string
[ "Returns", "Weed", "-", "FS", "master", "version" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/weed.py#L214-L225
train
ratt-ru/PyMORESANE
pymoresane/iuwt_convolution.py
fft_convolve
def fft_convolve(in1, in2, conv_device="cpu", conv_mode="linear", store_on_gpu=False): """ This function determines the convolution of two inputs using the FFT. Contains an implementation for both CPU and GPU. INPUTS: in1 (no default): Array containing one set of data, possibly an image. in2 (no default): Gpuarray containing the FFT of the PSF. conv_device (default = "cpu"): Parameter which allows specification of "cpu" or "gpu". conv_mode (default = "linear"): Mode specifier for the convolution - "linear" or "circular". """ # NOTE: Circular convolution assumes a periodic repetition of the input. This can cause edge effects. Linear # convolution pads the input with zeros to avoid this problem but is consequently heavier on computation and # memory. if conv_device=='gpu': if conv_mode=="linear": fft_in1 = pad_array(in1) fft_in1 = gpu_r2c_fft(fft_in1, store_on_gpu=True) fft_in2 = in2 conv_in1_in2 = fft_in1*fft_in2 conv_in1_in2 = contiguous_slice(fft_shift(gpu_c2r_ifft(conv_in1_in2, is_gpuarray=True, store_on_gpu=True))) if store_on_gpu: return conv_in1_in2 else: return conv_in1_in2.get() elif conv_mode=="circular": fft_in1 = gpu_r2c_fft(in1, store_on_gpu=True) fft_in2 = in2 conv_in1_in2 = fft_in1*fft_in2 conv_in1_in2 = fft_shift(gpu_c2r_ifft(conv_in1_in2, is_gpuarray=True, store_on_gpu=True)) if store_on_gpu: return conv_in1_in2 else: return conv_in1_in2.get() else: if conv_mode=="linear": fft_in1 = pad_array(in1) fft_in2 = in2 out1_slice = tuple(slice(0.5*sz,1.5*sz) for sz in in1.shape) return np.require(np.fft.fftshift(np.fft.irfft2(fft_in2*np.fft.rfft2(fft_in1)))[out1_slice], np.float32, 'C') elif conv_mode=="circular": return np.fft.fftshift(np.fft.irfft2(in2*np.fft.rfft2(in1)))
python
def fft_convolve(in1, in2, conv_device="cpu", conv_mode="linear", store_on_gpu=False): """ This function determines the convolution of two inputs using the FFT. Contains an implementation for both CPU and GPU. INPUTS: in1 (no default): Array containing one set of data, possibly an image. in2 (no default): Gpuarray containing the FFT of the PSF. conv_device (default = "cpu"): Parameter which allows specification of "cpu" or "gpu". conv_mode (default = "linear"): Mode specifier for the convolution - "linear" or "circular". """ # NOTE: Circular convolution assumes a periodic repetition of the input. This can cause edge effects. Linear # convolution pads the input with zeros to avoid this problem but is consequently heavier on computation and # memory. if conv_device=='gpu': if conv_mode=="linear": fft_in1 = pad_array(in1) fft_in1 = gpu_r2c_fft(fft_in1, store_on_gpu=True) fft_in2 = in2 conv_in1_in2 = fft_in1*fft_in2 conv_in1_in2 = contiguous_slice(fft_shift(gpu_c2r_ifft(conv_in1_in2, is_gpuarray=True, store_on_gpu=True))) if store_on_gpu: return conv_in1_in2 else: return conv_in1_in2.get() elif conv_mode=="circular": fft_in1 = gpu_r2c_fft(in1, store_on_gpu=True) fft_in2 = in2 conv_in1_in2 = fft_in1*fft_in2 conv_in1_in2 = fft_shift(gpu_c2r_ifft(conv_in1_in2, is_gpuarray=True, store_on_gpu=True)) if store_on_gpu: return conv_in1_in2 else: return conv_in1_in2.get() else: if conv_mode=="linear": fft_in1 = pad_array(in1) fft_in2 = in2 out1_slice = tuple(slice(0.5*sz,1.5*sz) for sz in in1.shape) return np.require(np.fft.fftshift(np.fft.irfft2(fft_in2*np.fft.rfft2(fft_in1)))[out1_slice], np.float32, 'C') elif conv_mode=="circular": return np.fft.fftshift(np.fft.irfft2(in2*np.fft.rfft2(in1)))
[ "def", "fft_convolve", "(", "in1", ",", "in2", ",", "conv_device", "=", "\"cpu\"", ",", "conv_mode", "=", "\"linear\"", ",", "store_on_gpu", "=", "False", ")", ":", "if", "conv_device", "==", "'gpu'", ":", "if", "conv_mode", "==", "\"linear\"", ":", "fft_in1", "=", "pad_array", "(", "in1", ")", "fft_in1", "=", "gpu_r2c_fft", "(", "fft_in1", ",", "store_on_gpu", "=", "True", ")", "fft_in2", "=", "in2", "conv_in1_in2", "=", "fft_in1", "*", "fft_in2", "conv_in1_in2", "=", "contiguous_slice", "(", "fft_shift", "(", "gpu_c2r_ifft", "(", "conv_in1_in2", ",", "is_gpuarray", "=", "True", ",", "store_on_gpu", "=", "True", ")", ")", ")", "if", "store_on_gpu", ":", "return", "conv_in1_in2", "else", ":", "return", "conv_in1_in2", ".", "get", "(", ")", "elif", "conv_mode", "==", "\"circular\"", ":", "fft_in1", "=", "gpu_r2c_fft", "(", "in1", ",", "store_on_gpu", "=", "True", ")", "fft_in2", "=", "in2", "conv_in1_in2", "=", "fft_in1", "*", "fft_in2", "conv_in1_in2", "=", "fft_shift", "(", "gpu_c2r_ifft", "(", "conv_in1_in2", ",", "is_gpuarray", "=", "True", ",", "store_on_gpu", "=", "True", ")", ")", "if", "store_on_gpu", ":", "return", "conv_in1_in2", "else", ":", "return", "conv_in1_in2", ".", "get", "(", ")", "else", ":", "if", "conv_mode", "==", "\"linear\"", ":", "fft_in1", "=", "pad_array", "(", "in1", ")", "fft_in2", "=", "in2", "out1_slice", "=", "tuple", "(", "slice", "(", "0.5", "*", "sz", ",", "1.5", "*", "sz", ")", "for", "sz", "in", "in1", ".", "shape", ")", "return", "np", ".", "require", "(", "np", ".", "fft", ".", "fftshift", "(", "np", ".", "fft", ".", "irfft2", "(", "fft_in2", "*", "np", ".", "fft", ".", "rfft2", "(", "fft_in1", ")", ")", ")", "[", "out1_slice", "]", ",", "np", ".", "float32", ",", "'C'", ")", "elif", "conv_mode", "==", "\"circular\"", ":", "return", "np", ".", "fft", ".", "fftshift", "(", "np", ".", "fft", ".", "irfft2", "(", "in2", "*", "np", ".", "fft", ".", "rfft2", "(", "in1", ")", ")", ")" ]
This function determines the convolution of two inputs using the FFT. Contains an implementation for both CPU and GPU. INPUTS: in1 (no default): Array containing one set of data, possibly an image. in2 (no default): Gpuarray containing the FFT of the PSF. conv_device (default = "cpu"): Parameter which allows specification of "cpu" or "gpu". conv_mode (default = "linear"): Mode specifier for the convolution - "linear" or "circular".
[ "This", "function", "determines", "the", "convolution", "of", "two", "inputs", "using", "the", "FFT", ".", "Contains", "an", "implementation", "for", "both", "CPU", "and", "GPU", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_convolution.py#L18-L73
train
ratt-ru/PyMORESANE
pymoresane/iuwt_convolution.py
gpu_r2c_fft
def gpu_r2c_fft(in1, is_gpuarray=False, store_on_gpu=False): """ This function makes use of the scikits implementation of the FFT for GPUs to take the real to complex FFT. INPUTS: in1 (no default): The array on which the FFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array. """ if is_gpuarray: gpu_in1 = in1 else: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.float32)) output_size = np.array(in1.shape) output_size[1] = 0.5*output_size[1] + 1 gpu_out1 = gpuarray.empty([output_size[0], output_size[1]], np.complex64) gpu_plan = Plan(gpu_in1.shape, np.float32, np.complex64) fft(gpu_in1, gpu_out1, gpu_plan) if store_on_gpu: return gpu_out1 else: return gpu_out1.get()
python
def gpu_r2c_fft(in1, is_gpuarray=False, store_on_gpu=False): """ This function makes use of the scikits implementation of the FFT for GPUs to take the real to complex FFT. INPUTS: in1 (no default): The array on which the FFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array. """ if is_gpuarray: gpu_in1 = in1 else: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.float32)) output_size = np.array(in1.shape) output_size[1] = 0.5*output_size[1] + 1 gpu_out1 = gpuarray.empty([output_size[0], output_size[1]], np.complex64) gpu_plan = Plan(gpu_in1.shape, np.float32, np.complex64) fft(gpu_in1, gpu_out1, gpu_plan) if store_on_gpu: return gpu_out1 else: return gpu_out1.get()
[ "def", "gpu_r2c_fft", "(", "in1", ",", "is_gpuarray", "=", "False", ",", "store_on_gpu", "=", "False", ")", ":", "if", "is_gpuarray", ":", "gpu_in1", "=", "in1", "else", ":", "gpu_in1", "=", "gpuarray", ".", "to_gpu_async", "(", "in1", ".", "astype", "(", "np", ".", "float32", ")", ")", "output_size", "=", "np", ".", "array", "(", "in1", ".", "shape", ")", "output_size", "[", "1", "]", "=", "0.5", "*", "output_size", "[", "1", "]", "+", "1", "gpu_out1", "=", "gpuarray", ".", "empty", "(", "[", "output_size", "[", "0", "]", ",", "output_size", "[", "1", "]", "]", ",", "np", ".", "complex64", ")", "gpu_plan", "=", "Plan", "(", "gpu_in1", ".", "shape", ",", "np", ".", "float32", ",", "np", ".", "complex64", ")", "fft", "(", "gpu_in1", ",", "gpu_out1", ",", "gpu_plan", ")", "if", "store_on_gpu", ":", "return", "gpu_out1", "else", ":", "return", "gpu_out1", ".", "get", "(", ")" ]
This function makes use of the scikits implementation of the FFT for GPUs to take the real to complex FFT. INPUTS: in1 (no default): The array on which the FFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array.
[ "This", "function", "makes", "use", "of", "the", "scikits", "implementation", "of", "the", "FFT", "for", "GPUs", "to", "take", "the", "real", "to", "complex", "FFT", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_convolution.py#L76-L106
train
ratt-ru/PyMORESANE
pymoresane/iuwt_convolution.py
gpu_c2r_ifft
def gpu_c2r_ifft(in1, is_gpuarray=False, store_on_gpu=False): """ This function makes use of the scikits implementation of the FFT for GPUs to take the complex to real IFFT. INPUTS: in1 (no default): The array on which the IFFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array. """ if is_gpuarray: gpu_in1 = in1 else: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.complex64)) output_size = np.array(in1.shape) output_size[1] = 2*(output_size[1]-1) gpu_out1 = gpuarray.empty([output_size[0],output_size[1]], np.float32) gpu_plan = Plan(output_size, np.complex64, np.float32) ifft(gpu_in1, gpu_out1, gpu_plan) scale_fft(gpu_out1) if store_on_gpu: return gpu_out1 else: return gpu_out1.get()
python
def gpu_c2r_ifft(in1, is_gpuarray=False, store_on_gpu=False): """ This function makes use of the scikits implementation of the FFT for GPUs to take the complex to real IFFT. INPUTS: in1 (no default): The array on which the IFFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array. """ if is_gpuarray: gpu_in1 = in1 else: gpu_in1 = gpuarray.to_gpu_async(in1.astype(np.complex64)) output_size = np.array(in1.shape) output_size[1] = 2*(output_size[1]-1) gpu_out1 = gpuarray.empty([output_size[0],output_size[1]], np.float32) gpu_plan = Plan(output_size, np.complex64, np.float32) ifft(gpu_in1, gpu_out1, gpu_plan) scale_fft(gpu_out1) if store_on_gpu: return gpu_out1 else: return gpu_out1.get()
[ "def", "gpu_c2r_ifft", "(", "in1", ",", "is_gpuarray", "=", "False", ",", "store_on_gpu", "=", "False", ")", ":", "if", "is_gpuarray", ":", "gpu_in1", "=", "in1", "else", ":", "gpu_in1", "=", "gpuarray", ".", "to_gpu_async", "(", "in1", ".", "astype", "(", "np", ".", "complex64", ")", ")", "output_size", "=", "np", ".", "array", "(", "in1", ".", "shape", ")", "output_size", "[", "1", "]", "=", "2", "*", "(", "output_size", "[", "1", "]", "-", "1", ")", "gpu_out1", "=", "gpuarray", ".", "empty", "(", "[", "output_size", "[", "0", "]", ",", "output_size", "[", "1", "]", "]", ",", "np", ".", "float32", ")", "gpu_plan", "=", "Plan", "(", "output_size", ",", "np", ".", "complex64", ",", "np", ".", "float32", ")", "ifft", "(", "gpu_in1", ",", "gpu_out1", ",", "gpu_plan", ")", "scale_fft", "(", "gpu_out1", ")", "if", "store_on_gpu", ":", "return", "gpu_out1", "else", ":", "return", "gpu_out1", ".", "get", "(", ")" ]
This function makes use of the scikits implementation of the FFT for GPUs to take the complex to real IFFT. INPUTS: in1 (no default): The array on which the IFFT is to be performed. is_gpuarray (default=True): Boolean specifier for whether or not input is on the gpu. store_on_gpu (default=False): Boolean specifier for whether the result is to be left on the gpu or not. OUTPUTS: gpu_out1 The gpu array containing the result. OR gpu_out1.get() The result from the gpu array.
[ "This", "function", "makes", "use", "of", "the", "scikits", "implementation", "of", "the", "FFT", "for", "GPUs", "to", "take", "the", "complex", "to", "real", "IFFT", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_convolution.py#L108-L139
train
ratt-ru/PyMORESANE
pymoresane/iuwt_convolution.py
pad_array
def pad_array(in1): """ Simple convenience function to pad arrays for linear convolution. INPUTS: in1 (no default): Input array which is to be padded. OUTPUTS: out1 Padded version of the input. """ padded_size = 2*np.array(in1.shape) out1 = np.zeros([padded_size[0],padded_size[1]]) out1[padded_size[0]/4:3*padded_size[0]/4,padded_size[1]/4:3*padded_size[1]/4] = in1 return out1
python
def pad_array(in1): """ Simple convenience function to pad arrays for linear convolution. INPUTS: in1 (no default): Input array which is to be padded. OUTPUTS: out1 Padded version of the input. """ padded_size = 2*np.array(in1.shape) out1 = np.zeros([padded_size[0],padded_size[1]]) out1[padded_size[0]/4:3*padded_size[0]/4,padded_size[1]/4:3*padded_size[1]/4] = in1 return out1
[ "def", "pad_array", "(", "in1", ")", ":", "padded_size", "=", "2", "*", "np", ".", "array", "(", "in1", ".", "shape", ")", "out1", "=", "np", ".", "zeros", "(", "[", "padded_size", "[", "0", "]", ",", "padded_size", "[", "1", "]", "]", ")", "out1", "[", "padded_size", "[", "0", "]", "/", "4", ":", "3", "*", "padded_size", "[", "0", "]", "/", "4", ",", "padded_size", "[", "1", "]", "/", "4", ":", "3", "*", "padded_size", "[", "1", "]", "/", "4", "]", "=", "in1", "return", "out1" ]
Simple convenience function to pad arrays for linear convolution. INPUTS: in1 (no default): Input array which is to be padded. OUTPUTS: out1 Padded version of the input.
[ "Simple", "convenience", "function", "to", "pad", "arrays", "for", "linear", "convolution", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/iuwt_convolution.py#L142-L158
train
brndnmtthws/dragon-rest
dragon_rest/dragons.py
DragonAPI.is_dragon
def is_dragon(host, timeout=1): """ Check if host is a dragon. Check if the specified host is a dragon based on simple heuristic. The code simply checks if particular strings are in the index page. It should work for DragonMint or Innosilicon branded miners. """ try: r = requests.get('http://{}/'.format(host), timeout=timeout) if r.status_code == 200: if '<title>DragonMint</title>' in r.text or \ '<title>AsicMiner</title>' in r.text: return True except requests.exceptions.RequestException: pass return False
python
def is_dragon(host, timeout=1): """ Check if host is a dragon. Check if the specified host is a dragon based on simple heuristic. The code simply checks if particular strings are in the index page. It should work for DragonMint or Innosilicon branded miners. """ try: r = requests.get('http://{}/'.format(host), timeout=timeout) if r.status_code == 200: if '<title>DragonMint</title>' in r.text or \ '<title>AsicMiner</title>' in r.text: return True except requests.exceptions.RequestException: pass return False
[ "def", "is_dragon", "(", "host", ",", "timeout", "=", "1", ")", ":", "try", ":", "r", "=", "requests", ".", "get", "(", "'http://{}/'", ".", "format", "(", "host", ")", ",", "timeout", "=", "timeout", ")", "if", "r", ".", "status_code", "==", "200", ":", "if", "'<title>DragonMint</title>'", "in", "r", ".", "text", "or", "'<title>AsicMiner</title>'", "in", "r", ".", "text", ":", "return", "True", "except", "requests", ".", "exceptions", ".", "RequestException", ":", "pass", "return", "False" ]
Check if host is a dragon. Check if the specified host is a dragon based on simple heuristic. The code simply checks if particular strings are in the index page. It should work for DragonMint or Innosilicon branded miners.
[ "Check", "if", "host", "is", "a", "dragon", "." ]
10ea09a6203c0cbfeeeb854702764bd778769887
https://github.com/brndnmtthws/dragon-rest/blob/10ea09a6203c0cbfeeeb854702764bd778769887/dragon_rest/dragons.py#L49-L65
train
brndnmtthws/dragon-rest
dragon_rest/dragons.py
DragonAPI.updatePools
def updatePools(self, pool1, username1, password1, pool2=None, username2=None, password2=None, pool3=None, username3=None, password3=None): """Change the pools of the miner. This call will restart cgminer.""" return self.__post('/api/updatePools', data={ 'Pool1': pool1, 'UserName1': username1, 'Password1': password1, 'Pool2': pool2, 'UserName2': username2, 'Password2': password2, 'Pool3': pool3, 'UserName3': username3, 'Password3': password3, })
python
def updatePools(self, pool1, username1, password1, pool2=None, username2=None, password2=None, pool3=None, username3=None, password3=None): """Change the pools of the miner. This call will restart cgminer.""" return self.__post('/api/updatePools', data={ 'Pool1': pool1, 'UserName1': username1, 'Password1': password1, 'Pool2': pool2, 'UserName2': username2, 'Password2': password2, 'Pool3': pool3, 'UserName3': username3, 'Password3': password3, })
[ "def", "updatePools", "(", "self", ",", "pool1", ",", "username1", ",", "password1", ",", "pool2", "=", "None", ",", "username2", "=", "None", ",", "password2", "=", "None", ",", "pool3", "=", "None", ",", "username3", "=", "None", ",", "password3", "=", "None", ")", ":", "return", "self", ".", "__post", "(", "'/api/updatePools'", ",", "data", "=", "{", "'Pool1'", ":", "pool1", ",", "'UserName1'", ":", "username1", ",", "'Password1'", ":", "password1", ",", "'Pool2'", ":", "pool2", ",", "'UserName2'", ":", "username2", ",", "'Password2'", ":", "password2", ",", "'Pool3'", ":", "pool3", ",", "'UserName3'", ":", "username3", ",", "'Password3'", ":", "password3", ",", "}", ")" ]
Change the pools of the miner. This call will restart cgminer.
[ "Change", "the", "pools", "of", "the", "miner", ".", "This", "call", "will", "restart", "cgminer", "." ]
10ea09a6203c0cbfeeeb854702764bd778769887
https://github.com/brndnmtthws/dragon-rest/blob/10ea09a6203c0cbfeeeb854702764bd778769887/dragon_rest/dragons.py#L148-L170
train
brndnmtthws/dragon-rest
dragon_rest/dragons.py
DragonAPI.updatePassword
def updatePassword(self, user, currentPassword, newPassword): """Change the password of a user.""" return self.__post('/api/updatePassword', data={ 'user': user, 'currentPassword': currentPassword, 'newPassword': newPassword })
python
def updatePassword(self, user, currentPassword, newPassword): """Change the password of a user.""" return self.__post('/api/updatePassword', data={ 'user': user, 'currentPassword': currentPassword, 'newPassword': newPassword })
[ "def", "updatePassword", "(", "self", ",", "user", ",", "currentPassword", ",", "newPassword", ")", ":", "return", "self", ".", "__post", "(", "'/api/updatePassword'", ",", "data", "=", "{", "'user'", ":", "user", ",", "'currentPassword'", ":", "currentPassword", ",", "'newPassword'", ":", "newPassword", "}", ")" ]
Change the password of a user.
[ "Change", "the", "password", "of", "a", "user", "." ]
10ea09a6203c0cbfeeeb854702764bd778769887
https://github.com/brndnmtthws/dragon-rest/blob/10ea09a6203c0cbfeeeb854702764bd778769887/dragon_rest/dragons.py#L172-L182
train
brndnmtthws/dragon-rest
dragon_rest/dragons.py
DragonAPI.updateNetwork
def updateNetwork(self, dhcp='dhcp', ipaddress=None, netmask=None, gateway=None, dns=None): """Change the current network settings.""" return self.__post('/api/updateNetwork', data={ 'dhcp': dhcp, 'ipaddress': ipaddress, 'netmask': netmask, 'gateway': gateway, 'dns': json.dumps(dns) })
python
def updateNetwork(self, dhcp='dhcp', ipaddress=None, netmask=None, gateway=None, dns=None): """Change the current network settings.""" return self.__post('/api/updateNetwork', data={ 'dhcp': dhcp, 'ipaddress': ipaddress, 'netmask': netmask, 'gateway': gateway, 'dns': json.dumps(dns) })
[ "def", "updateNetwork", "(", "self", ",", "dhcp", "=", "'dhcp'", ",", "ipaddress", "=", "None", ",", "netmask", "=", "None", ",", "gateway", "=", "None", ",", "dns", "=", "None", ")", ":", "return", "self", ".", "__post", "(", "'/api/updateNetwork'", ",", "data", "=", "{", "'dhcp'", ":", "dhcp", ",", "'ipaddress'", ":", "ipaddress", ",", "'netmask'", ":", "netmask", ",", "'gateway'", ":", "gateway", ",", "'dns'", ":", "json", ".", "dumps", "(", "dns", ")", "}", ")" ]
Change the current network settings.
[ "Change", "the", "current", "network", "settings", "." ]
10ea09a6203c0cbfeeeb854702764bd778769887
https://github.com/brndnmtthws/dragon-rest/blob/10ea09a6203c0cbfeeeb854702764bd778769887/dragon_rest/dragons.py#L188-L202
train
brndnmtthws/dragon-rest
dragon_rest/dragons.py
DragonAPI.upgradeUpload
def upgradeUpload(self, file): """Upgrade the firmware of the miner.""" files = {'upfile': open(file, 'rb')} return self.__post_files('/upgrade/upload', files=files)
python
def upgradeUpload(self, file): """Upgrade the firmware of the miner.""" files = {'upfile': open(file, 'rb')} return self.__post_files('/upgrade/upload', files=files)
[ "def", "upgradeUpload", "(", "self", ",", "file", ")", ":", "files", "=", "{", "'upfile'", ":", "open", "(", "file", ",", "'rb'", ")", "}", "return", "self", ".", "__post_files", "(", "'/upgrade/upload'", ",", "files", "=", "files", ")" ]
Upgrade the firmware of the miner.
[ "Upgrade", "the", "firmware", "of", "the", "miner", "." ]
10ea09a6203c0cbfeeeb854702764bd778769887
https://github.com/brndnmtthws/dragon-rest/blob/10ea09a6203c0cbfeeeb854702764bd778769887/dragon_rest/dragons.py#L237-L241
train
tuomas2/automate
src/automate/statusobject.py
StatusObject.is_program
def is_program(self): """ A property which can be used to check if StatusObject uses program features or not. """ from automate.callables import Empty return not (isinstance(self.on_activate, Empty) and isinstance(self.on_deactivate, Empty) and isinstance(self.on_update, Empty))
python
def is_program(self): """ A property which can be used to check if StatusObject uses program features or not. """ from automate.callables import Empty return not (isinstance(self.on_activate, Empty) and isinstance(self.on_deactivate, Empty) and isinstance(self.on_update, Empty))
[ "def", "is_program", "(", "self", ")", ":", "from", "automate", ".", "callables", "import", "Empty", "return", "not", "(", "isinstance", "(", "self", ".", "on_activate", ",", "Empty", ")", "and", "isinstance", "(", "self", ".", "on_deactivate", ",", "Empty", ")", "and", "isinstance", "(", "self", ".", "on_update", ",", "Empty", ")", ")" ]
A property which can be used to check if StatusObject uses program features or not.
[ "A", "property", "which", "can", "be", "used", "to", "check", "if", "StatusObject", "uses", "program", "features", "or", "not", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/statusobject.py#L208-L215
train
tuomas2/automate
src/automate/statusobject.py
StatusObject.get_as_datadict
def get_as_datadict(self): """ Get data of this object as a data dictionary. Used by websocket service. """ d = super().get_as_datadict() d.update(dict(status=self.status, data_type=self.data_type, editable=self.editable)) return d
python
def get_as_datadict(self): """ Get data of this object as a data dictionary. Used by websocket service. """ d = super().get_as_datadict() d.update(dict(status=self.status, data_type=self.data_type, editable=self.editable)) return d
[ "def", "get_as_datadict", "(", "self", ")", ":", "d", "=", "super", "(", ")", ".", "get_as_datadict", "(", ")", "d", ".", "update", "(", "dict", "(", "status", "=", "self", ".", "status", ",", "data_type", "=", "self", ".", "data_type", ",", "editable", "=", "self", ".", "editable", ")", ")", "return", "d" ]
Get data of this object as a data dictionary. Used by websocket service.
[ "Get", "data", "of", "this", "object", "as", "a", "data", "dictionary", ".", "Used", "by", "websocket", "service", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/statusobject.py#L238-L244
train
tuomas2/automate
src/automate/statusobject.py
StatusObject._do_change_status
def _do_change_status(self, status, force=False): """ This function is called by - set_status - _update_program_stack if active program is being changed - thia may be launched by sensor status change. status lock is necessary because these happen from different threads. This does not directly change status, but adds change request to queue. """ self.system.worker_thread.put(DummyStatusWorkerTask(self._request_status_change_in_queue, status, force=force))
python
def _do_change_status(self, status, force=False): """ This function is called by - set_status - _update_program_stack if active program is being changed - thia may be launched by sensor status change. status lock is necessary because these happen from different threads. This does not directly change status, but adds change request to queue. """ self.system.worker_thread.put(DummyStatusWorkerTask(self._request_status_change_in_queue, status, force=force))
[ "def", "_do_change_status", "(", "self", ",", "status", ",", "force", "=", "False", ")", ":", "self", ".", "system", ".", "worker_thread", ".", "put", "(", "DummyStatusWorkerTask", "(", "self", ".", "_request_status_change_in_queue", ",", "status", ",", "force", "=", "force", ")", ")" ]
This function is called by - set_status - _update_program_stack if active program is being changed - thia may be launched by sensor status change. status lock is necessary because these happen from different threads. This does not directly change status, but adds change request to queue.
[ "This", "function", "is", "called", "by", "-", "set_status", "-", "_update_program_stack", "if", "active", "program", "is", "being", "changed", "-", "thia", "may", "be", "launched", "by", "sensor", "status", "change", ".", "status", "lock", "is", "necessary", "because", "these", "happen", "from", "different", "threads", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/statusobject.py#L354-L366
train
tuomas2/automate
src/automate/statusobject.py
AbstractActuator.activate_program
def activate_program(self, program): """ Called by program which desires to manipulate this actuator, when it is activated. """ self.logger.debug("activate_program %s", program) if program in self.program_stack: return with self._program_lock: self.logger.debug("activate_program got through %s", program) self.program_stack.append(program) self._update_program_stack()
python
def activate_program(self, program): """ Called by program which desires to manipulate this actuator, when it is activated. """ self.logger.debug("activate_program %s", program) if program in self.program_stack: return with self._program_lock: self.logger.debug("activate_program got through %s", program) self.program_stack.append(program) self._update_program_stack()
[ "def", "activate_program", "(", "self", ",", "program", ")", ":", "self", ".", "logger", ".", "debug", "(", "\"activate_program %s\"", ",", "program", ")", "if", "program", "in", "self", ".", "program_stack", ":", "return", "with", "self", ".", "_program_lock", ":", "self", ".", "logger", ".", "debug", "(", "\"activate_program got through %s\"", ",", "program", ")", "self", ".", "program_stack", ".", "append", "(", "program", ")", "self", ".", "_update_program_stack", "(", ")" ]
Called by program which desires to manipulate this actuator, when it is activated.
[ "Called", "by", "program", "which", "desires", "to", "manipulate", "this", "actuator", "when", "it", "is", "activated", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/statusobject.py#L578-L589
train
tuomas2/automate
src/automate/statusobject.py
AbstractActuator.deactivate_program
def deactivate_program(self, program): """ Called by program, when it is deactivated. """ self.logger.debug("deactivate_program %s", program) with self._program_lock: self.logger.debug("deactivate_program got through %s", program) if program not in self.program_stack: import ipdb ipdb.set_trace() self.program_stack.remove(program) if program in self.program_status: del self.program_status[program] self._update_program_stack()
python
def deactivate_program(self, program): """ Called by program, when it is deactivated. """ self.logger.debug("deactivate_program %s", program) with self._program_lock: self.logger.debug("deactivate_program got through %s", program) if program not in self.program_stack: import ipdb ipdb.set_trace() self.program_stack.remove(program) if program in self.program_status: del self.program_status[program] self._update_program_stack()
[ "def", "deactivate_program", "(", "self", ",", "program", ")", ":", "self", ".", "logger", ".", "debug", "(", "\"deactivate_program %s\"", ",", "program", ")", "with", "self", ".", "_program_lock", ":", "self", ".", "logger", ".", "debug", "(", "\"deactivate_program got through %s\"", ",", "program", ")", "if", "program", "not", "in", "self", ".", "program_stack", ":", "import", "ipdb", "ipdb", ".", "set_trace", "(", ")", "self", ".", "program_stack", ".", "remove", "(", "program", ")", "if", "program", "in", "self", ".", "program_status", ":", "del", "self", ".", "program_status", "[", "program", "]", "self", ".", "_update_program_stack", "(", ")" ]
Called by program, when it is deactivated.
[ "Called", "by", "program", "when", "it", "is", "deactivated", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/statusobject.py#L591-L605
train
praekeltfoundation/seaworthy
seaworthy/stream/logs.py
stream_logs
def stream_logs(container, timeout=10.0, **logs_kwargs): """ Stream logs from a Docker container within a timeout. :param ~docker.models.containers.Container container: Container who's log lines to stream. :param timeout: Timeout value in seconds. :param logs_kwargs: Additional keyword arguments to pass to ``container.logs()``. For example, the ``stdout`` and ``stderr`` boolean arguments can be used to determine whether to stream stdout or stderr or both (the default). :raises TimeoutError: When the timeout value is reached before the logs have completed. """ stream = container.logs(stream=True, **logs_kwargs) return stream_timeout( stream, timeout, 'Timeout waiting for container logs.')
python
def stream_logs(container, timeout=10.0, **logs_kwargs): """ Stream logs from a Docker container within a timeout. :param ~docker.models.containers.Container container: Container who's log lines to stream. :param timeout: Timeout value in seconds. :param logs_kwargs: Additional keyword arguments to pass to ``container.logs()``. For example, the ``stdout`` and ``stderr`` boolean arguments can be used to determine whether to stream stdout or stderr or both (the default). :raises TimeoutError: When the timeout value is reached before the logs have completed. """ stream = container.logs(stream=True, **logs_kwargs) return stream_timeout( stream, timeout, 'Timeout waiting for container logs.')
[ "def", "stream_logs", "(", "container", ",", "timeout", "=", "10.0", ",", "**", "logs_kwargs", ")", ":", "stream", "=", "container", ".", "logs", "(", "stream", "=", "True", ",", "**", "logs_kwargs", ")", "return", "stream_timeout", "(", "stream", ",", "timeout", ",", "'Timeout waiting for container logs.'", ")" ]
Stream logs from a Docker container within a timeout. :param ~docker.models.containers.Container container: Container who's log lines to stream. :param timeout: Timeout value in seconds. :param logs_kwargs: Additional keyword arguments to pass to ``container.logs()``. For example, the ``stdout`` and ``stderr`` boolean arguments can be used to determine whether to stream stdout or stderr or both (the default). :raises TimeoutError: When the timeout value is reached before the logs have completed.
[ "Stream", "logs", "from", "a", "Docker", "container", "within", "a", "timeout", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/stream/logs.py#L8-L26
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
fetch_image
def fetch_image(client, name): """ Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name. """ try: image = client.images.get(name) except docker.errors.ImageNotFound: name, tag = _parse_image_tag(name) tag = 'latest' if tag is None else tag log.info("Pulling tag '{}' for image '{}'...".format(tag, name)) image = client.images.pull(name, tag=tag) log.debug("Found image '{}' for tag '{}'".format(image.id, name)) return image
python
def fetch_image(client, name): """ Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name. """ try: image = client.images.get(name) except docker.errors.ImageNotFound: name, tag = _parse_image_tag(name) tag = 'latest' if tag is None else tag log.info("Pulling tag '{}' for image '{}'...".format(tag, name)) image = client.images.pull(name, tag=tag) log.debug("Found image '{}' for tag '{}'".format(image.id, name)) return image
[ "def", "fetch_image", "(", "client", ",", "name", ")", ":", "try", ":", "image", "=", "client", ".", "images", ".", "get", "(", "name", ")", "except", "docker", ".", "errors", ".", "ImageNotFound", ":", "name", ",", "tag", "=", "_parse_image_tag", "(", "name", ")", "tag", "=", "'latest'", "if", "tag", "is", "None", "else", "tag", "log", ".", "info", "(", "\"Pulling tag '{}' for image '{}'...\"", ".", "format", "(", "tag", ",", "name", ")", ")", "image", "=", "client", ".", "images", ".", "pull", "(", "name", ",", "tag", "=", "tag", ")", "log", ".", "debug", "(", "\"Found image '{}' for tag '{}'\"", ".", "format", "(", "image", ".", "id", ",", "name", ")", ")", "return", "image" ]
Fetch an image if it isn't already present. This works like ``docker pull`` and will pull the tag ``latest`` if no tag is specified in the image name.
[ "Fetch", "an", "image", "if", "it", "isn", "t", "already", "present", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L27-L44
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
_HelperBase._get_id_and_model
def _get_id_and_model(self, id_or_model): """ Get both the model and ID of an object that could be an ID or a model. :param id_or_model: The object that could be an ID string or a model object. :param model_collection: The collection to which the model belongs. """ if isinstance(id_or_model, self.collection.model): model = id_or_model elif isinstance(id_or_model, str): # Assume we have an ID string model = self.collection.get(id_or_model) else: raise TypeError('Unexpected type {}, expected {} or {}'.format( type(id_or_model), str, self.collection.model)) return model.id, model
python
def _get_id_and_model(self, id_or_model): """ Get both the model and ID of an object that could be an ID or a model. :param id_or_model: The object that could be an ID string or a model object. :param model_collection: The collection to which the model belongs. """ if isinstance(id_or_model, self.collection.model): model = id_or_model elif isinstance(id_or_model, str): # Assume we have an ID string model = self.collection.get(id_or_model) else: raise TypeError('Unexpected type {}, expected {} or {}'.format( type(id_or_model), str, self.collection.model)) return model.id, model
[ "def", "_get_id_and_model", "(", "self", ",", "id_or_model", ")", ":", "if", "isinstance", "(", "id_or_model", ",", "self", ".", "collection", ".", "model", ")", ":", "model", "=", "id_or_model", "elif", "isinstance", "(", "id_or_model", ",", "str", ")", ":", "model", "=", "self", ".", "collection", ".", "get", "(", "id_or_model", ")", "else", ":", "raise", "TypeError", "(", "'Unexpected type {}, expected {} or {}'", ".", "format", "(", "type", "(", "id_or_model", ")", ",", "str", ",", "self", ".", "collection", ".", "model", ")", ")", "return", "model", ".", "id", ",", "model" ]
Get both the model and ID of an object that could be an ID or a model. :param id_or_model: The object that could be an ID string or a model object. :param model_collection: The collection to which the model belongs.
[ "Get", "both", "the", "model", "and", "ID", "of", "an", "object", "that", "could", "be", "an", "ID", "or", "a", "model", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L82-L100
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
_HelperBase.create
def create(self, name, *args, **kwargs): """ Create an instance of this resource type. """ resource_name = self._resource_name(name) log.info( "Creating {} '{}'...".format(self._model_name, resource_name)) resource = self.collection.create(*args, name=resource_name, **kwargs) self._ids.add(resource.id) return resource
python
def create(self, name, *args, **kwargs): """ Create an instance of this resource type. """ resource_name = self._resource_name(name) log.info( "Creating {} '{}'...".format(self._model_name, resource_name)) resource = self.collection.create(*args, name=resource_name, **kwargs) self._ids.add(resource.id) return resource
[ "def", "create", "(", "self", ",", "name", ",", "*", "args", ",", "**", "kwargs", ")", ":", "resource_name", "=", "self", ".", "_resource_name", "(", "name", ")", "log", ".", "info", "(", "\"Creating {} '{}'...\"", ".", "format", "(", "self", ".", "_model_name", ",", "resource_name", ")", ")", "resource", "=", "self", ".", "collection", ".", "create", "(", "*", "args", ",", "name", "=", "resource_name", ",", "**", "kwargs", ")", "self", ".", "_ids", ".", "add", "(", "resource", ".", "id", ")", "return", "resource" ]
Create an instance of this resource type.
[ "Create", "an", "instance", "of", "this", "resource", "type", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L102-L111
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
_HelperBase.remove
def remove(self, resource, **kwargs): """ Remove an instance of this resource type. """ log.info( "Removing {} '{}'...".format(self._model_name, resource.name)) resource.remove(**kwargs) self._ids.remove(resource.id)
python
def remove(self, resource, **kwargs): """ Remove an instance of this resource type. """ log.info( "Removing {} '{}'...".format(self._model_name, resource.name)) resource.remove(**kwargs) self._ids.remove(resource.id)
[ "def", "remove", "(", "self", ",", "resource", ",", "**", "kwargs", ")", ":", "log", ".", "info", "(", "\"Removing {} '{}'...\"", ".", "format", "(", "self", ".", "_model_name", ",", "resource", ".", "name", ")", ")", "resource", ".", "remove", "(", "**", "kwargs", ")", "self", ".", "_ids", ".", "remove", "(", "resource", ".", "id", ")" ]
Remove an instance of this resource type.
[ "Remove", "an", "instance", "of", "this", "resource", "type", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L113-L120
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
ContainerHelper.remove
def remove(self, container, force=True, volumes=True): """ Remove a container. :param container: The container to remove. :param force: Whether to force the removal of the container, even if it is running. Note that this defaults to True, unlike the Docker default. :param volumes: Whether to remove any volumes that were created implicitly with this container, i.e. any volumes that were created due to ``VOLUME`` directives in the Dockerfile. External volumes that were manually created will not be removed. Note that this defaults to True, unlike the Docker default (where the equivalent parameter, ``v``, defaults to False). """ super().remove(container, force=force, v=volumes)
python
def remove(self, container, force=True, volumes=True): """ Remove a container. :param container: The container to remove. :param force: Whether to force the removal of the container, even if it is running. Note that this defaults to True, unlike the Docker default. :param volumes: Whether to remove any volumes that were created implicitly with this container, i.e. any volumes that were created due to ``VOLUME`` directives in the Dockerfile. External volumes that were manually created will not be removed. Note that this defaults to True, unlike the Docker default (where the equivalent parameter, ``v``, defaults to False). """ super().remove(container, force=force, v=volumes)
[ "def", "remove", "(", "self", ",", "container", ",", "force", "=", "True", ",", "volumes", "=", "True", ")", ":", "super", "(", ")", ".", "remove", "(", "container", ",", "force", "=", "force", ",", "v", "=", "volumes", ")" ]
Remove a container. :param container: The container to remove. :param force: Whether to force the removal of the container, even if it is running. Note that this defaults to True, unlike the Docker default. :param volumes: Whether to remove any volumes that were created implicitly with this container, i.e. any volumes that were created due to ``VOLUME`` directives in the Dockerfile. External volumes that were manually created will not be removed. Note that this defaults to True, unlike the Docker default (where the equivalent parameter, ``v``, defaults to False).
[ "Remove", "a", "container", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L266-L283
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
NetworkHelper.get_default
def get_default(self, create=True): """ Get the default bridge network that containers are connected to if no other network options are specified. :param create: Whether or not to create the network if it doesn't already exist. """ if self._default_network is None and create: log.debug("Creating default network...") self._default_network = self.create('default', driver='bridge') return self._default_network
python
def get_default(self, create=True): """ Get the default bridge network that containers are connected to if no other network options are specified. :param create: Whether or not to create the network if it doesn't already exist. """ if self._default_network is None and create: log.debug("Creating default network...") self._default_network = self.create('default', driver='bridge') return self._default_network
[ "def", "get_default", "(", "self", ",", "create", "=", "True", ")", ":", "if", "self", ".", "_default_network", "is", "None", "and", "create", ":", "log", ".", "debug", "(", "\"Creating default network...\"", ")", "self", ".", "_default_network", "=", "self", ".", "create", "(", "'default'", ",", "driver", "=", "'bridge'", ")", "return", "self", ".", "_default_network" ]
Get the default bridge network that containers are connected to if no other network options are specified. :param create: Whether or not to create the network if it doesn't already exist.
[ "Get", "the", "default", "bridge", "network", "that", "containers", "are", "connected", "to", "if", "no", "other", "network", "options", "are", "specified", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L326-L338
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
DockerHelper._helper_for_model
def _helper_for_model(self, model_type): """ Get the helper for a given type of Docker model. For use by resource definitions. """ if model_type is models.containers.Container: return self.containers if model_type is models.images.Image: return self.images if model_type is models.networks.Network: return self.networks if model_type is models.volumes.Volume: return self.volumes raise ValueError('Unknown model type {}'.format(model_type))
python
def _helper_for_model(self, model_type): """ Get the helper for a given type of Docker model. For use by resource definitions. """ if model_type is models.containers.Container: return self.containers if model_type is models.images.Image: return self.images if model_type is models.networks.Network: return self.networks if model_type is models.volumes.Volume: return self.volumes raise ValueError('Unknown model type {}'.format(model_type))
[ "def", "_helper_for_model", "(", "self", ",", "model_type", ")", ":", "if", "model_type", "is", "models", ".", "containers", ".", "Container", ":", "return", "self", ".", "containers", "if", "model_type", "is", "models", ".", "images", ".", "Image", ":", "return", "self", ".", "images", "if", "model_type", "is", "models", ".", "networks", ".", "Network", ":", "return", "self", ".", "networks", "if", "model_type", "is", "models", ".", "volumes", ".", "Volume", ":", "return", "self", ".", "volumes", "raise", "ValueError", "(", "'Unknown model type {}'", ".", "format", "(", "model_type", ")", ")" ]
Get the helper for a given type of Docker model. For use by resource definitions.
[ "Get", "the", "helper", "for", "a", "given", "type", "of", "Docker", "model", ".", "For", "use", "by", "resource", "definitions", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L398-L412
train
praekeltfoundation/seaworthy
seaworthy/helpers.py
DockerHelper.teardown
def teardown(self): """ Clean up all resources when we're done with them. """ self.containers._teardown() self.networks._teardown() self.volumes._teardown() # We need to close the underlying APIClient explicitly to avoid # ResourceWarnings from unclosed HTTP connections. self._client.api.close()
python
def teardown(self): """ Clean up all resources when we're done with them. """ self.containers._teardown() self.networks._teardown() self.volumes._teardown() # We need to close the underlying APIClient explicitly to avoid # ResourceWarnings from unclosed HTTP connections. self._client.api.close()
[ "def", "teardown", "(", "self", ")", ":", "self", ".", "containers", ".", "_teardown", "(", ")", "self", ".", "networks", ".", "_teardown", "(", ")", "self", ".", "volumes", ".", "_teardown", "(", ")", "self", ".", "_client", ".", "api", ".", "close", "(", ")" ]
Clean up all resources when we're done with them.
[ "Clean", "up", "all", "resources", "when", "we", "re", "done", "with", "them", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/helpers.py#L414-L424
train
praekeltfoundation/seaworthy
seaworthy/containers/redis.py
RedisContainer.exec_redis_cli
def exec_redis_cli(self, command, args=[], db=0, redis_cli_opts=[]): """ Execute a ``redis-cli`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param db: the db number to query (default ``0``) :param redis_cli_opts: a list of extra options to pass to ``redis-cli`` :returns: a tuple of the command exit code and output """ cli_opts = ['-n', str(db)] + redis_cli_opts cmd = ['redis-cli'] + cli_opts + [command] + [str(a) for a in args] return self.inner().exec_run(cmd)
python
def exec_redis_cli(self, command, args=[], db=0, redis_cli_opts=[]): """ Execute a ``redis-cli`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param db: the db number to query (default ``0``) :param redis_cli_opts: a list of extra options to pass to ``redis-cli`` :returns: a tuple of the command exit code and output """ cli_opts = ['-n', str(db)] + redis_cli_opts cmd = ['redis-cli'] + cli_opts + [command] + [str(a) for a in args] return self.inner().exec_run(cmd)
[ "def", "exec_redis_cli", "(", "self", ",", "command", ",", "args", "=", "[", "]", ",", "db", "=", "0", ",", "redis_cli_opts", "=", "[", "]", ")", ":", "cli_opts", "=", "[", "'-n'", ",", "str", "(", "db", ")", "]", "+", "redis_cli_opts", "cmd", "=", "[", "'redis-cli'", "]", "+", "cli_opts", "+", "[", "command", "]", "+", "[", "str", "(", "a", ")", "for", "a", "in", "args", "]", "return", "self", ".", "inner", "(", ")", ".", "exec_run", "(", "cmd", ")" ]
Execute a ``redis-cli`` command inside a running container. :param command: the command to run :param args: a list of args for the command :param db: the db number to query (default ``0``) :param redis_cli_opts: a list of extra options to pass to ``redis-cli`` :returns: a tuple of the command exit code and output
[ "Execute", "a", "redis", "-", "cli", "command", "inside", "a", "running", "container", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/redis.py#L40-L52
train
praekeltfoundation/seaworthy
seaworthy/containers/redis.py
RedisContainer.list_keys
def list_keys(self, pattern='*', db=0): """ Run the ``KEYS`` command and return the list of matching keys. :param pattern: the pattern to filter keys by (default ``*``) :param db: the db number to query (default ``0``) """ lines = output_lines(self.exec_redis_cli('KEYS', [pattern], db=db)) return [] if lines == [''] else lines
python
def list_keys(self, pattern='*', db=0): """ Run the ``KEYS`` command and return the list of matching keys. :param pattern: the pattern to filter keys by (default ``*``) :param db: the db number to query (default ``0``) """ lines = output_lines(self.exec_redis_cli('KEYS', [pattern], db=db)) return [] if lines == [''] else lines
[ "def", "list_keys", "(", "self", ",", "pattern", "=", "'*'", ",", "db", "=", "0", ")", ":", "lines", "=", "output_lines", "(", "self", ".", "exec_redis_cli", "(", "'KEYS'", ",", "[", "pattern", "]", ",", "db", "=", "db", ")", ")", "return", "[", "]", "if", "lines", "==", "[", "''", "]", "else", "lines" ]
Run the ``KEYS`` command and return the list of matching keys. :param pattern: the pattern to filter keys by (default ``*``) :param db: the db number to query (default ``0``)
[ "Run", "the", "KEYS", "command", "and", "return", "the", "list", "of", "matching", "keys", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/containers/redis.py#L54-L62
train
tuomas2/automate
src/automate/common.py
threaded
def threaded(system, func, *args, **kwargs): """ uses thread_init as a decorator-style """ @wraps(func) def wrapper(*args, **kwargs): try: return func(*args, **kwargs) except Exception as e: if system.raven_client: system.raven_client.captureException() logger.exception('Exception occurred in thread: %s', e) return False return lambda: wrapper(*args, **kwargs)
python
def threaded(system, func, *args, **kwargs): """ uses thread_init as a decorator-style """ @wraps(func) def wrapper(*args, **kwargs): try: return func(*args, **kwargs) except Exception as e: if system.raven_client: system.raven_client.captureException() logger.exception('Exception occurred in thread: %s', e) return False return lambda: wrapper(*args, **kwargs)
[ "def", "threaded", "(", "system", ",", "func", ",", "*", "args", ",", "**", "kwargs", ")", ":", "@", "wraps", "(", "func", ")", "def", "wrapper", "(", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "return", "func", "(", "*", "args", ",", "**", "kwargs", ")", "except", "Exception", "as", "e", ":", "if", "system", ".", "raven_client", ":", "system", ".", "raven_client", ".", "captureException", "(", ")", "logger", ".", "exception", "(", "'Exception occurred in thread: %s'", ",", "e", ")", "return", "False", "return", "lambda", ":", "wrapper", "(", "*", "args", ",", "**", "kwargs", ")" ]
uses thread_init as a decorator-style
[ "uses", "thread_init", "as", "a", "decorator", "-", "style" ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/common.py#L96-L108
train
praekeltfoundation/seaworthy
seaworthy/stream/matchers.py
OrderedMatcher.match
def match(self, item): """ Return ``True`` if the expected matchers are matched in the expected order, otherwise ``False``. """ if self._position == len(self._matchers): raise RuntimeError('Matcher exhausted, no more matchers to use') matcher = self._matchers[self._position] if matcher(item): self._position += 1 if self._position == len(self._matchers): # All patterns have been matched return True return False
python
def match(self, item): """ Return ``True`` if the expected matchers are matched in the expected order, otherwise ``False``. """ if self._position == len(self._matchers): raise RuntimeError('Matcher exhausted, no more matchers to use') matcher = self._matchers[self._position] if matcher(item): self._position += 1 if self._position == len(self._matchers): # All patterns have been matched return True return False
[ "def", "match", "(", "self", ",", "item", ")", ":", "if", "self", ".", "_position", "==", "len", "(", "self", ".", "_matchers", ")", ":", "raise", "RuntimeError", "(", "'Matcher exhausted, no more matchers to use'", ")", "matcher", "=", "self", ".", "_matchers", "[", "self", ".", "_position", "]", "if", "matcher", "(", "item", ")", ":", "self", ".", "_position", "+=", "1", "if", "self", ".", "_position", "==", "len", "(", "self", ".", "_matchers", ")", ":", "return", "True", "return", "False" ]
Return ``True`` if the expected matchers are matched in the expected order, otherwise ``False``.
[ "Return", "True", "if", "the", "expected", "matchers", "are", "matched", "in", "the", "expected", "order", "otherwise", "False", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/stream/matchers.py#L81-L97
train
praekeltfoundation/seaworthy
seaworthy/stream/matchers.py
UnorderedMatcher.match
def match(self, item): """ Return ``True`` if the expected matchers are matched in any order, otherwise ``False``. """ if not self._unused_matchers: raise RuntimeError('Matcher exhausted, no more matchers to use') for matcher in self._unused_matchers: if matcher(item): self._used_matchers.append(matcher) break if not self._unused_matchers: # All patterns have been matched return True return False
python
def match(self, item): """ Return ``True`` if the expected matchers are matched in any order, otherwise ``False``. """ if not self._unused_matchers: raise RuntimeError('Matcher exhausted, no more matchers to use') for matcher in self._unused_matchers: if matcher(item): self._used_matchers.append(matcher) break if not self._unused_matchers: # All patterns have been matched return True return False
[ "def", "match", "(", "self", ",", "item", ")", ":", "if", "not", "self", ".", "_unused_matchers", ":", "raise", "RuntimeError", "(", "'Matcher exhausted, no more matchers to use'", ")", "for", "matcher", "in", "self", ".", "_unused_matchers", ":", "if", "matcher", "(", "item", ")", ":", "self", ".", "_used_matchers", ".", "append", "(", "matcher", ")", "break", "if", "not", "self", ".", "_unused_matchers", ":", "return", "True", "return", "False" ]
Return ``True`` if the expected matchers are matched in any order, otherwise ``False``.
[ "Return", "True", "if", "the", "expected", "matchers", "are", "matched", "in", "any", "order", "otherwise", "False", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/stream/matchers.py#L129-L146
train
ratt-ru/PyMORESANE
pymoresane/main.py
DataImage.moresane_by_scale
def moresane_by_scale(self, start_scale=1, stop_scale=20, subregion=None, sigma_level=4, loop_gain=0.1, tolerance=0.75, accuracy=1e-6, major_loop_miter=100, minor_loop_miter=30, all_on_gpu=False, decom_mode="ser", core_count=1, conv_device='cpu', conv_mode='linear', extraction_mode='cpu', enforce_positivity=False, edge_suppression=False, edge_offset=0, flux_threshold=0, neg_comp=False, edge_excl=0, int_excl=0): """ Extension of the MORESANE algorithm. This takes a scale-by-scale approach, attempting to remove all sources at the lower scales before moving onto the higher ones. At each step the algorithm may return to previous scales to remove the sources uncovered by the deconvolution. INPUTS: start_scale (default=1) The first scale which is to be considered. stop_scale (default=20) The maximum scale which is to be considered. Optional. subregion (default=None): Size, in pixels, of the central region to be analyzed and deconvolved. sigma_level (default=4) Number of sigma at which thresholding is to be performed. loop_gain (default=0.1): Loop gain for the deconvolution. tolerance (default=0.75): Tolerance level for object extraction. Significant objects contain wavelet coefficients greater than the tolerance multiplied by the maximum wavelet coefficient in the scale under consideration. accuracy (default=1e-6): Threshold on the standard deviation of the residual noise. Exit main loop when this threshold is reached. major_loop_miter (default=100): Maximum number of iterations allowed in the major loop. Exit condition. minor_loop_miter (default=30): Maximum number of iterations allowed in the minor loop. Serves as an exit condition when the SNR does not reach a maximum. all_on_gpu (default=False): Boolean specifier to toggle all gpu modes on. decom_mode (default='ser'): Specifier for decomposition mode - serial, multiprocessing, or gpu. core_count (default=1): In the event that multiprocessing, specifies the number of cores. conv_device (default='cpu'): Specifier for device to be used - cpu or gpu. conv_mode (default='linear'): Specifier for convolution mode - linear or circular. extraction_mode (default='cpu'): Specifier for mode to be used - cpu or gpu. enforce_positivity (default=False): Boolean specifier for whether or not a model must be strictly positive. edge_suppression (default=False): Boolean specifier for whether or not the edges are to be suprressed. edge_offset (default=0): Numeric value for an additional user-specified number of edge pixels to be ignored. This is added to the minimum suppression. OUTPUTS: self.model (no default): Model extracted by the algorithm. self.residual (no default): Residual signal after deconvolution. """ # The following preserves the dirty image as it will be changed on every iteration. dirty_data = self.dirty_data scale_count = start_scale while not (self.complete): logger.info("MORESANE at scale {}".format(scale_count)) self.moresane(subregion=subregion, scale_count=scale_count, sigma_level=sigma_level, loop_gain=loop_gain, tolerance=tolerance, accuracy=accuracy, major_loop_miter=major_loop_miter, minor_loop_miter=minor_loop_miter, all_on_gpu=all_on_gpu, decom_mode=decom_mode, core_count=core_count, conv_device=conv_device, conv_mode=conv_mode, extraction_mode=extraction_mode, enforce_positivity=enforce_positivity, edge_suppression=edge_suppression, edge_offset=edge_offset, flux_threshold=flux_threshold, neg_comp=neg_comp, edge_excl=edge_excl, int_excl=int_excl) self.dirty_data = self.residual scale_count += 1 if (scale_count>(np.log2(self.dirty_data.shape[0]))-1): logger.info("Maximum scale reached - finished.") break if (scale_count>stop_scale): logger.info("Maximum scale reached - finished.") break # Restores the original dirty image. self.dirty_data = dirty_data self.complete = False
python
def moresane_by_scale(self, start_scale=1, stop_scale=20, subregion=None, sigma_level=4, loop_gain=0.1, tolerance=0.75, accuracy=1e-6, major_loop_miter=100, minor_loop_miter=30, all_on_gpu=False, decom_mode="ser", core_count=1, conv_device='cpu', conv_mode='linear', extraction_mode='cpu', enforce_positivity=False, edge_suppression=False, edge_offset=0, flux_threshold=0, neg_comp=False, edge_excl=0, int_excl=0): """ Extension of the MORESANE algorithm. This takes a scale-by-scale approach, attempting to remove all sources at the lower scales before moving onto the higher ones. At each step the algorithm may return to previous scales to remove the sources uncovered by the deconvolution. INPUTS: start_scale (default=1) The first scale which is to be considered. stop_scale (default=20) The maximum scale which is to be considered. Optional. subregion (default=None): Size, in pixels, of the central region to be analyzed and deconvolved. sigma_level (default=4) Number of sigma at which thresholding is to be performed. loop_gain (default=0.1): Loop gain for the deconvolution. tolerance (default=0.75): Tolerance level for object extraction. Significant objects contain wavelet coefficients greater than the tolerance multiplied by the maximum wavelet coefficient in the scale under consideration. accuracy (default=1e-6): Threshold on the standard deviation of the residual noise. Exit main loop when this threshold is reached. major_loop_miter (default=100): Maximum number of iterations allowed in the major loop. Exit condition. minor_loop_miter (default=30): Maximum number of iterations allowed in the minor loop. Serves as an exit condition when the SNR does not reach a maximum. all_on_gpu (default=False): Boolean specifier to toggle all gpu modes on. decom_mode (default='ser'): Specifier for decomposition mode - serial, multiprocessing, or gpu. core_count (default=1): In the event that multiprocessing, specifies the number of cores. conv_device (default='cpu'): Specifier for device to be used - cpu or gpu. conv_mode (default='linear'): Specifier for convolution mode - linear or circular. extraction_mode (default='cpu'): Specifier for mode to be used - cpu or gpu. enforce_positivity (default=False): Boolean specifier for whether or not a model must be strictly positive. edge_suppression (default=False): Boolean specifier for whether or not the edges are to be suprressed. edge_offset (default=0): Numeric value for an additional user-specified number of edge pixels to be ignored. This is added to the minimum suppression. OUTPUTS: self.model (no default): Model extracted by the algorithm. self.residual (no default): Residual signal after deconvolution. """ # The following preserves the dirty image as it will be changed on every iteration. dirty_data = self.dirty_data scale_count = start_scale while not (self.complete): logger.info("MORESANE at scale {}".format(scale_count)) self.moresane(subregion=subregion, scale_count=scale_count, sigma_level=sigma_level, loop_gain=loop_gain, tolerance=tolerance, accuracy=accuracy, major_loop_miter=major_loop_miter, minor_loop_miter=minor_loop_miter, all_on_gpu=all_on_gpu, decom_mode=decom_mode, core_count=core_count, conv_device=conv_device, conv_mode=conv_mode, extraction_mode=extraction_mode, enforce_positivity=enforce_positivity, edge_suppression=edge_suppression, edge_offset=edge_offset, flux_threshold=flux_threshold, neg_comp=neg_comp, edge_excl=edge_excl, int_excl=int_excl) self.dirty_data = self.residual scale_count += 1 if (scale_count>(np.log2(self.dirty_data.shape[0]))-1): logger.info("Maximum scale reached - finished.") break if (scale_count>stop_scale): logger.info("Maximum scale reached - finished.") break # Restores the original dirty image. self.dirty_data = dirty_data self.complete = False
[ "def", "moresane_by_scale", "(", "self", ",", "start_scale", "=", "1", ",", "stop_scale", "=", "20", ",", "subregion", "=", "None", ",", "sigma_level", "=", "4", ",", "loop_gain", "=", "0.1", ",", "tolerance", "=", "0.75", ",", "accuracy", "=", "1e-6", ",", "major_loop_miter", "=", "100", ",", "minor_loop_miter", "=", "30", ",", "all_on_gpu", "=", "False", ",", "decom_mode", "=", "\"ser\"", ",", "core_count", "=", "1", ",", "conv_device", "=", "'cpu'", ",", "conv_mode", "=", "'linear'", ",", "extraction_mode", "=", "'cpu'", ",", "enforce_positivity", "=", "False", ",", "edge_suppression", "=", "False", ",", "edge_offset", "=", "0", ",", "flux_threshold", "=", "0", ",", "neg_comp", "=", "False", ",", "edge_excl", "=", "0", ",", "int_excl", "=", "0", ")", ":", "dirty_data", "=", "self", ".", "dirty_data", "scale_count", "=", "start_scale", "while", "not", "(", "self", ".", "complete", ")", ":", "logger", ".", "info", "(", "\"MORESANE at scale {}\"", ".", "format", "(", "scale_count", ")", ")", "self", ".", "moresane", "(", "subregion", "=", "subregion", ",", "scale_count", "=", "scale_count", ",", "sigma_level", "=", "sigma_level", ",", "loop_gain", "=", "loop_gain", ",", "tolerance", "=", "tolerance", ",", "accuracy", "=", "accuracy", ",", "major_loop_miter", "=", "major_loop_miter", ",", "minor_loop_miter", "=", "minor_loop_miter", ",", "all_on_gpu", "=", "all_on_gpu", ",", "decom_mode", "=", "decom_mode", ",", "core_count", "=", "core_count", ",", "conv_device", "=", "conv_device", ",", "conv_mode", "=", "conv_mode", ",", "extraction_mode", "=", "extraction_mode", ",", "enforce_positivity", "=", "enforce_positivity", ",", "edge_suppression", "=", "edge_suppression", ",", "edge_offset", "=", "edge_offset", ",", "flux_threshold", "=", "flux_threshold", ",", "neg_comp", "=", "neg_comp", ",", "edge_excl", "=", "edge_excl", ",", "int_excl", "=", "int_excl", ")", "self", ".", "dirty_data", "=", "self", ".", "residual", "scale_count", "+=", "1", "if", "(", "scale_count", ">", "(", "np", ".", "log2", "(", "self", ".", "dirty_data", ".", "shape", "[", "0", "]", ")", ")", "-", "1", ")", ":", "logger", ".", "info", "(", "\"Maximum scale reached - finished.\"", ")", "break", "if", "(", "scale_count", ">", "stop_scale", ")", ":", "logger", ".", "info", "(", "\"Maximum scale reached - finished.\"", ")", "break", "self", ".", "dirty_data", "=", "dirty_data", "self", ".", "complete", "=", "False" ]
Extension of the MORESANE algorithm. This takes a scale-by-scale approach, attempting to remove all sources at the lower scales before moving onto the higher ones. At each step the algorithm may return to previous scales to remove the sources uncovered by the deconvolution. INPUTS: start_scale (default=1) The first scale which is to be considered. stop_scale (default=20) The maximum scale which is to be considered. Optional. subregion (default=None): Size, in pixels, of the central region to be analyzed and deconvolved. sigma_level (default=4) Number of sigma at which thresholding is to be performed. loop_gain (default=0.1): Loop gain for the deconvolution. tolerance (default=0.75): Tolerance level for object extraction. Significant objects contain wavelet coefficients greater than the tolerance multiplied by the maximum wavelet coefficient in the scale under consideration. accuracy (default=1e-6): Threshold on the standard deviation of the residual noise. Exit main loop when this threshold is reached. major_loop_miter (default=100): Maximum number of iterations allowed in the major loop. Exit condition. minor_loop_miter (default=30): Maximum number of iterations allowed in the minor loop. Serves as an exit condition when the SNR does not reach a maximum. all_on_gpu (default=False): Boolean specifier to toggle all gpu modes on. decom_mode (default='ser'): Specifier for decomposition mode - serial, multiprocessing, or gpu. core_count (default=1): In the event that multiprocessing, specifies the number of cores. conv_device (default='cpu'): Specifier for device to be used - cpu or gpu. conv_mode (default='linear'): Specifier for convolution mode - linear or circular. extraction_mode (default='cpu'): Specifier for mode to be used - cpu or gpu. enforce_positivity (default=False): Boolean specifier for whether or not a model must be strictly positive. edge_suppression (default=False): Boolean specifier for whether or not the edges are to be suprressed. edge_offset (default=0): Numeric value for an additional user-specified number of edge pixels to be ignored. This is added to the minimum suppression. OUTPUTS: self.model (no default): Model extracted by the algorithm. self.residual (no default): Residual signal after deconvolution.
[ "Extension", "of", "the", "MORESANE", "algorithm", ".", "This", "takes", "a", "scale", "-", "by", "-", "scale", "approach", "attempting", "to", "remove", "all", "sources", "at", "the", "lower", "scales", "before", "moving", "onto", "the", "higher", "ones", ".", "At", "each", "step", "the", "algorithm", "may", "return", "to", "previous", "scales", "to", "remove", "the", "sources", "uncovered", "by", "the", "deconvolution", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/main.py#L523-L599
train
ratt-ru/PyMORESANE
pymoresane/main.py
DataImage.restore
def restore(self): """ This method constructs the restoring beam and then adds the convolution to the residual. """ clean_beam, beam_params = beam_fit(self.psf_data, self.cdelt1, self.cdelt2) if np.all(np.array(self.psf_data_shape)==2*np.array(self.dirty_data_shape)): self.restored = np.fft.fftshift(np.fft.irfft2(np.fft.rfft2(conv.pad_array(self.model))*np.fft.rfft2(clean_beam))) self.restored = self.restored[self.dirty_data_shape[0]/2:-self.dirty_data_shape[0]/2, self.dirty_data_shape[1]/2:-self.dirty_data_shape[1]/2] else: self.restored = np.fft.fftshift(np.fft.irfft2(np.fft.rfft2(self.model)*np.fft.rfft2(clean_beam))) self.restored += self.residual self.restored = self.restored.astype(np.float32) return beam_params
python
def restore(self): """ This method constructs the restoring beam and then adds the convolution to the residual. """ clean_beam, beam_params = beam_fit(self.psf_data, self.cdelt1, self.cdelt2) if np.all(np.array(self.psf_data_shape)==2*np.array(self.dirty_data_shape)): self.restored = np.fft.fftshift(np.fft.irfft2(np.fft.rfft2(conv.pad_array(self.model))*np.fft.rfft2(clean_beam))) self.restored = self.restored[self.dirty_data_shape[0]/2:-self.dirty_data_shape[0]/2, self.dirty_data_shape[1]/2:-self.dirty_data_shape[1]/2] else: self.restored = np.fft.fftshift(np.fft.irfft2(np.fft.rfft2(self.model)*np.fft.rfft2(clean_beam))) self.restored += self.residual self.restored = self.restored.astype(np.float32) return beam_params
[ "def", "restore", "(", "self", ")", ":", "clean_beam", ",", "beam_params", "=", "beam_fit", "(", "self", ".", "psf_data", ",", "self", ".", "cdelt1", ",", "self", ".", "cdelt2", ")", "if", "np", ".", "all", "(", "np", ".", "array", "(", "self", ".", "psf_data_shape", ")", "==", "2", "*", "np", ".", "array", "(", "self", ".", "dirty_data_shape", ")", ")", ":", "self", ".", "restored", "=", "np", ".", "fft", ".", "fftshift", "(", "np", ".", "fft", ".", "irfft2", "(", "np", ".", "fft", ".", "rfft2", "(", "conv", ".", "pad_array", "(", "self", ".", "model", ")", ")", "*", "np", ".", "fft", ".", "rfft2", "(", "clean_beam", ")", ")", ")", "self", ".", "restored", "=", "self", ".", "restored", "[", "self", ".", "dirty_data_shape", "[", "0", "]", "/", "2", ":", "-", "self", ".", "dirty_data_shape", "[", "0", "]", "/", "2", ",", "self", ".", "dirty_data_shape", "[", "1", "]", "/", "2", ":", "-", "self", ".", "dirty_data_shape", "[", "1", "]", "/", "2", "]", "else", ":", "self", ".", "restored", "=", "np", ".", "fft", ".", "fftshift", "(", "np", ".", "fft", ".", "irfft2", "(", "np", ".", "fft", ".", "rfft2", "(", "self", ".", "model", ")", "*", "np", ".", "fft", ".", "rfft2", "(", "clean_beam", ")", ")", ")", "self", ".", "restored", "+=", "self", ".", "residual", "self", ".", "restored", "=", "self", ".", "restored", ".", "astype", "(", "np", ".", "float32", ")", "return", "beam_params" ]
This method constructs the restoring beam and then adds the convolution to the residual.
[ "This", "method", "constructs", "the", "restoring", "beam", "and", "then", "adds", "the", "convolution", "to", "the", "residual", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/main.py#L601-L616
train
ratt-ru/PyMORESANE
pymoresane/main.py
DataImage.handle_input
def handle_input(self, input_hdr): """ This method tries to ensure that the input data has the correct dimensions. INPUTS: input_hdr (no default) Header from which data shape is to be extracted. """ input_slice = input_hdr['NAXIS']*[0] for i in range(input_hdr['NAXIS']): if input_hdr['CTYPE%d'%(i+1)].startswith("RA"): input_slice[-1] = slice(None) if input_hdr['CTYPE%d'%(i+1)].startswith("DEC"): input_slice[-2] = slice(None) return input_slice
python
def handle_input(self, input_hdr): """ This method tries to ensure that the input data has the correct dimensions. INPUTS: input_hdr (no default) Header from which data shape is to be extracted. """ input_slice = input_hdr['NAXIS']*[0] for i in range(input_hdr['NAXIS']): if input_hdr['CTYPE%d'%(i+1)].startswith("RA"): input_slice[-1] = slice(None) if input_hdr['CTYPE%d'%(i+1)].startswith("DEC"): input_slice[-2] = slice(None) return input_slice
[ "def", "handle_input", "(", "self", ",", "input_hdr", ")", ":", "input_slice", "=", "input_hdr", "[", "'NAXIS'", "]", "*", "[", "0", "]", "for", "i", "in", "range", "(", "input_hdr", "[", "'NAXIS'", "]", ")", ":", "if", "input_hdr", "[", "'CTYPE%d'", "%", "(", "i", "+", "1", ")", "]", ".", "startswith", "(", "\"RA\"", ")", ":", "input_slice", "[", "-", "1", "]", "=", "slice", "(", "None", ")", "if", "input_hdr", "[", "'CTYPE%d'", "%", "(", "i", "+", "1", ")", "]", ".", "startswith", "(", "\"DEC\"", ")", ":", "input_slice", "[", "-", "2", "]", "=", "slice", "(", "None", ")", "return", "input_slice" ]
This method tries to ensure that the input data has the correct dimensions. INPUTS: input_hdr (no default) Header from which data shape is to be extracted.
[ "This", "method", "tries", "to", "ensure", "that", "the", "input", "data", "has", "the", "correct", "dimensions", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/main.py#L618-L634
train
ratt-ru/PyMORESANE
pymoresane/main.py
DataImage.save_fits
def save_fits(self, data, name): """ This method simply saves the model components and the residual. INPUTS: data (no default) Data which is to be saved. name (no default) File name for new .fits file. Will overwrite. """ data = data.reshape(1, 1, data.shape[0], data.shape[0]) new_file = pyfits.PrimaryHDU(data,self.img_hdu_list[0].header) new_file.writeto("{}".format(name), overwrite=True)
python
def save_fits(self, data, name): """ This method simply saves the model components and the residual. INPUTS: data (no default) Data which is to be saved. name (no default) File name for new .fits file. Will overwrite. """ data = data.reshape(1, 1, data.shape[0], data.shape[0]) new_file = pyfits.PrimaryHDU(data,self.img_hdu_list[0].header) new_file.writeto("{}".format(name), overwrite=True)
[ "def", "save_fits", "(", "self", ",", "data", ",", "name", ")", ":", "data", "=", "data", ".", "reshape", "(", "1", ",", "1", ",", "data", ".", "shape", "[", "0", "]", ",", "data", ".", "shape", "[", "0", "]", ")", "new_file", "=", "pyfits", ".", "PrimaryHDU", "(", "data", ",", "self", ".", "img_hdu_list", "[", "0", "]", ".", "header", ")", "new_file", ".", "writeto", "(", "\"{}\"", ".", "format", "(", "name", ")", ",", "overwrite", "=", "True", ")" ]
This method simply saves the model components and the residual. INPUTS: data (no default) Data which is to be saved. name (no default) File name for new .fits file. Will overwrite.
[ "This", "method", "simply", "saves", "the", "model", "components", "and", "the", "residual", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/main.py#L636-L646
train
ratt-ru/PyMORESANE
pymoresane/main.py
DataImage.make_logger
def make_logger(self, level="INFO"): """ Convenience function which creates a logger for the module. INPUTS: level (default="INFO"): Minimum log level for logged/streamed messages. OUTPUTS: logger Logger for the function. NOTE: Must be bound to variable named logger. """ level = getattr(logging, level.upper()) logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) fh = logging.FileHandler('PyMORESANE.log', mode='w') fh.setLevel(level) ch = logging.StreamHandler() ch.setLevel(level) formatter = logging.Formatter('%(asctime)s [%(levelname)s]: %(''message)s', datefmt='[%m/%d/%Y] [%I:%M:%S]') fh.setFormatter(formatter) ch.setFormatter(formatter) logger.addHandler(fh) logger.addHandler(ch) return logger
python
def make_logger(self, level="INFO"): """ Convenience function which creates a logger for the module. INPUTS: level (default="INFO"): Minimum log level for logged/streamed messages. OUTPUTS: logger Logger for the function. NOTE: Must be bound to variable named logger. """ level = getattr(logging, level.upper()) logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) fh = logging.FileHandler('PyMORESANE.log', mode='w') fh.setLevel(level) ch = logging.StreamHandler() ch.setLevel(level) formatter = logging.Formatter('%(asctime)s [%(levelname)s]: %(''message)s', datefmt='[%m/%d/%Y] [%I:%M:%S]') fh.setFormatter(formatter) ch.setFormatter(formatter) logger.addHandler(fh) logger.addHandler(ch) return logger
[ "def", "make_logger", "(", "self", ",", "level", "=", "\"INFO\"", ")", ":", "level", "=", "getattr", "(", "logging", ",", "level", ".", "upper", "(", ")", ")", "logger", "=", "logging", ".", "getLogger", "(", "__name__", ")", "logger", ".", "setLevel", "(", "logging", ".", "DEBUG", ")", "fh", "=", "logging", ".", "FileHandler", "(", "'PyMORESANE.log'", ",", "mode", "=", "'w'", ")", "fh", ".", "setLevel", "(", "level", ")", "ch", "=", "logging", ".", "StreamHandler", "(", ")", "ch", ".", "setLevel", "(", "level", ")", "formatter", "=", "logging", ".", "Formatter", "(", "'%(asctime)s [%(levelname)s]: %('", "'message)s'", ",", "datefmt", "=", "'[%m/%d/%Y] [%I:%M:%S]'", ")", "fh", ".", "setFormatter", "(", "formatter", ")", "ch", ".", "setFormatter", "(", "formatter", ")", "logger", ".", "addHandler", "(", "fh", ")", "logger", ".", "addHandler", "(", "ch", ")", "return", "logger" ]
Convenience function which creates a logger for the module. INPUTS: level (default="INFO"): Minimum log level for logged/streamed messages. OUTPUTS: logger Logger for the function. NOTE: Must be bound to variable named logger.
[ "Convenience", "function", "which", "creates", "a", "logger", "for", "the", "module", "." ]
b024591ad0bbb69320d08841f28a2c27f62ae1af
https://github.com/ratt-ru/PyMORESANE/blob/b024591ad0bbb69320d08841f28a2c27f62ae1af/pymoresane/main.py#L648-L676
train
tuomas2/automate
src/automate/services/textui.py
TextUIService.text_ui
def text_ui(self): """ Start Text UI main loop """ self.logger.info("Starting command line interface") self.help() try: self.ipython_ui() except ImportError: self.fallback_ui() self.system.cleanup()
python
def text_ui(self): """ Start Text UI main loop """ self.logger.info("Starting command line interface") self.help() try: self.ipython_ui() except ImportError: self.fallback_ui() self.system.cleanup()
[ "def", "text_ui", "(", "self", ")", ":", "self", ".", "logger", ".", "info", "(", "\"Starting command line interface\"", ")", "self", ".", "help", "(", ")", "try", ":", "self", ".", "ipython_ui", "(", ")", "except", "ImportError", ":", "self", ".", "fallback_ui", "(", ")", "self", ".", "system", ".", "cleanup", "(", ")" ]
Start Text UI main loop
[ "Start", "Text", "UI", "main", "loop" ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/services/textui.py#L96-L106
train
utek/pyseaweed
pyseaweed/utils.py
Connection._prepare_headers
def _prepare_headers(self, additional_headers=None, **kwargs): """Prepare headers for http communication. Return dict of header to be used in requests. Args: .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Headers dict. Key and values are string """ user_agent = "pyseaweed/{version}".format(version=__version__) headers = {"User-Agent": user_agent} if additional_headers is not None: headers.update(additional_headers) return headers
python
def _prepare_headers(self, additional_headers=None, **kwargs): """Prepare headers for http communication. Return dict of header to be used in requests. Args: .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Headers dict. Key and values are string """ user_agent = "pyseaweed/{version}".format(version=__version__) headers = {"User-Agent": user_agent} if additional_headers is not None: headers.update(additional_headers) return headers
[ "def", "_prepare_headers", "(", "self", ",", "additional_headers", "=", "None", ",", "**", "kwargs", ")", ":", "user_agent", "=", "\"pyseaweed/{version}\"", ".", "format", "(", "version", "=", "__version__", ")", "headers", "=", "{", "\"User-Agent\"", ":", "user_agent", "}", "if", "additional_headers", "is", "not", "None", ":", "headers", ".", "update", "(", "additional_headers", ")", "return", "headers" ]
Prepare headers for http communication. Return dict of header to be used in requests. Args: .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Headers dict. Key and values are string
[ "Prepare", "headers", "for", "http", "communication", "." ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L21-L39
train
utek/pyseaweed
pyseaweed/utils.py
Connection.head
def head(self, url, *args, **kwargs): """Returns response to http HEAD on provided url """ res = self._conn.head(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res return None
python
def head(self, url, *args, **kwargs): """Returns response to http HEAD on provided url """ res = self._conn.head(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res return None
[ "def", "head", "(", "self", ",", "url", ",", "*", "args", ",", "**", "kwargs", ")", ":", "res", "=", "self", ".", "_conn", ".", "head", "(", "url", ",", "headers", "=", "self", ".", "_prepare_headers", "(", "**", "kwargs", ")", ")", "if", "res", ".", "status_code", "==", "200", ":", "return", "res", "return", "None" ]
Returns response to http HEAD on provided url
[ "Returns", "response", "to", "http", "HEAD", "on", "provided", "url" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L41-L48
train
utek/pyseaweed
pyseaweed/utils.py
Connection.get_data
def get_data(self, url, *args, **kwargs): """Gets data from url as text Returns content under the provided url as text Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string """ res = self._conn.get(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res.text else: return None
python
def get_data(self, url, *args, **kwargs): """Gets data from url as text Returns content under the provided url as text Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string """ res = self._conn.get(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res.text else: return None
[ "def", "get_data", "(", "self", ",", "url", ",", "*", "args", ",", "**", "kwargs", ")", ":", "res", "=", "self", ".", "_conn", ".", "get", "(", "url", ",", "headers", "=", "self", ".", "_prepare_headers", "(", "**", "kwargs", ")", ")", "if", "res", ".", "status_code", "==", "200", ":", "return", "res", ".", "text", "else", ":", "return", "None" ]
Gets data from url as text Returns content under the provided url as text Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string
[ "Gets", "data", "from", "url", "as", "text" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L50-L70
train
utek/pyseaweed
pyseaweed/utils.py
Connection.get_raw_data
def get_raw_data(self, url, *args, **kwargs): """Gets data from url as bytes Returns content under the provided url as bytes ie. for binary data Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: bytes """ res = self._conn.get(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res.content else: return None
python
def get_raw_data(self, url, *args, **kwargs): """Gets data from url as bytes Returns content under the provided url as bytes ie. for binary data Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: bytes """ res = self._conn.get(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200: return res.content else: return None
[ "def", "get_raw_data", "(", "self", ",", "url", ",", "*", "args", ",", "**", "kwargs", ")", ":", "res", "=", "self", ".", "_conn", ".", "get", "(", "url", ",", "headers", "=", "self", ".", "_prepare_headers", "(", "**", "kwargs", ")", ")", "if", "res", ".", "status_code", "==", "200", ":", "return", "res", ".", "content", "else", ":", "return", "None" ]
Gets data from url as bytes Returns content under the provided url as bytes ie. for binary data Args: **url**: address of the wanted data .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: bytes
[ "Gets", "data", "from", "url", "as", "bytes" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L72-L93
train
utek/pyseaweed
pyseaweed/utils.py
Connection.post_file
def post_file(self, url, filename, file_stream, *args, **kwargs): """Uploads file to provided url. Returns contents as text Args: **url**: address where to upload file **filename**: Name of the uploaded file **file_stream**: file like object to upload .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string """ res = self._conn.post(url, files={filename: file_stream}, headers=self._prepare_headers(**kwargs)) if res.status_code == 200 or res.status_code == 201: return res.text else: return None
python
def post_file(self, url, filename, file_stream, *args, **kwargs): """Uploads file to provided url. Returns contents as text Args: **url**: address where to upload file **filename**: Name of the uploaded file **file_stream**: file like object to upload .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string """ res = self._conn.post(url, files={filename: file_stream}, headers=self._prepare_headers(**kwargs)) if res.status_code == 200 or res.status_code == 201: return res.text else: return None
[ "def", "post_file", "(", "self", ",", "url", ",", "filename", ",", "file_stream", ",", "*", "args", ",", "**", "kwargs", ")", ":", "res", "=", "self", ".", "_conn", ".", "post", "(", "url", ",", "files", "=", "{", "filename", ":", "file_stream", "}", ",", "headers", "=", "self", ".", "_prepare_headers", "(", "**", "kwargs", ")", ")", "if", "res", ".", "status_code", "==", "200", "or", "res", ".", "status_code", "==", "201", ":", "return", "res", ".", "text", "else", ":", "return", "None" ]
Uploads file to provided url. Returns contents as text Args: **url**: address where to upload file **filename**: Name of the uploaded file **file_stream**: file like object to upload .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: string
[ "Uploads", "file", "to", "provided", "url", "." ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L95-L119
train
utek/pyseaweed
pyseaweed/utils.py
Connection.delete_data
def delete_data(self, url, *args, **kwargs): """Deletes data under provided url Returns status as boolean. Args: **url**: address of file to be deleted .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Boolean. True if request was successful. False if not. """ res = self._conn.delete(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200 or res.status_code == 202: return True else: return False
python
def delete_data(self, url, *args, **kwargs): """Deletes data under provided url Returns status as boolean. Args: **url**: address of file to be deleted .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Boolean. True if request was successful. False if not. """ res = self._conn.delete(url, headers=self._prepare_headers(**kwargs)) if res.status_code == 200 or res.status_code == 202: return True else: return False
[ "def", "delete_data", "(", "self", ",", "url", ",", "*", "args", ",", "**", "kwargs", ")", ":", "res", "=", "self", ".", "_conn", ".", "delete", "(", "url", ",", "headers", "=", "self", ".", "_prepare_headers", "(", "**", "kwargs", ")", ")", "if", "res", ".", "status_code", "==", "200", "or", "res", ".", "status_code", "==", "202", ":", "return", "True", "else", ":", "return", "False" ]
Deletes data under provided url Returns status as boolean. Args: **url**: address of file to be deleted .. versionadded:: 0.3.2 **additional_headers**: (optional) Additional headers to be used with request Returns: Boolean. True if request was successful. False if not.
[ "Deletes", "data", "under", "provided", "url" ]
218049329885425a2b8370157fa44952e64516be
https://github.com/utek/pyseaweed/blob/218049329885425a2b8370157fa44952e64516be/pyseaweed/utils.py#L121-L140
train
jtauber/greek-accentuation
greek_accentuation/characters.py
remove_diacritic
def remove_diacritic(*diacritics): """ Given a collection of Unicode diacritics, return a function that takes a string and returns the string without those diacritics. """ def _(text): return unicodedata.normalize("NFC", "".join( ch for ch in unicodedata.normalize("NFD", text) if ch not in diacritics) ) return _
python
def remove_diacritic(*diacritics): """ Given a collection of Unicode diacritics, return a function that takes a string and returns the string without those diacritics. """ def _(text): return unicodedata.normalize("NFC", "".join( ch for ch in unicodedata.normalize("NFD", text) if ch not in diacritics) ) return _
[ "def", "remove_diacritic", "(", "*", "diacritics", ")", ":", "def", "_", "(", "text", ")", ":", "return", "unicodedata", ".", "normalize", "(", "\"NFC\"", ",", "\"\"", ".", "join", "(", "ch", "for", "ch", "in", "unicodedata", ".", "normalize", "(", "\"NFD\"", ",", "text", ")", "if", "ch", "not", "in", "diacritics", ")", ")", "return", "_" ]
Given a collection of Unicode diacritics, return a function that takes a string and returns the string without those diacritics.
[ "Given", "a", "collection", "of", "Unicode", "diacritics", "return", "a", "function", "that", "takes", "a", "string", "and", "returns", "the", "string", "without", "those", "diacritics", "." ]
330796cd97f7c7adcbecbd05bd91be984f9b9f67
https://github.com/jtauber/greek-accentuation/blob/330796cd97f7c7adcbecbd05bd91be984f9b9f67/greek_accentuation/characters.py#L42-L53
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
deep_merge
def deep_merge(*dicts): """ Recursively merge all input dicts into a single dict. """ result = {} for d in dicts: if not isinstance(d, dict): raise Exception('Can only deep_merge dicts, got {}'.format(d)) for k, v in d.items(): # Whenever the value is a dict, we deep_merge it. This ensures that # (a) we only ever merge dicts with dicts and (b) we always get a # deep(ish) copy of the dicts and are thus safe from accidental # mutations to shared state. if isinstance(v, dict): v = deep_merge(result.get(k, {}), v) result[k] = v return result
python
def deep_merge(*dicts): """ Recursively merge all input dicts into a single dict. """ result = {} for d in dicts: if not isinstance(d, dict): raise Exception('Can only deep_merge dicts, got {}'.format(d)) for k, v in d.items(): # Whenever the value is a dict, we deep_merge it. This ensures that # (a) we only ever merge dicts with dicts and (b) we always get a # deep(ish) copy of the dicts and are thus safe from accidental # mutations to shared state. if isinstance(v, dict): v = deep_merge(result.get(k, {}), v) result[k] = v return result
[ "def", "deep_merge", "(", "*", "dicts", ")", ":", "result", "=", "{", "}", "for", "d", "in", "dicts", ":", "if", "not", "isinstance", "(", "d", ",", "dict", ")", ":", "raise", "Exception", "(", "'Can only deep_merge dicts, got {}'", ".", "format", "(", "d", ")", ")", "for", "k", ",", "v", "in", "d", ".", "items", "(", ")", ":", "if", "isinstance", "(", "v", ",", "dict", ")", ":", "v", "=", "deep_merge", "(", "result", ".", "get", "(", "k", ",", "{", "}", ")", ",", "v", ")", "result", "[", "k", "]", "=", "v", "return", "result" ]
Recursively merge all input dicts into a single dict.
[ "Recursively", "merge", "all", "input", "dicts", "into", "a", "single", "dict", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L21-L37
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
_DefinitionBase.create
def create(self, **kwargs): """ Create an instance of this resource definition. Only one instance may exist at any given time. """ if self.created: raise RuntimeError( '{} already created.'.format(self.__model_type__.__name__)) kwargs = self.merge_kwargs(self._create_kwargs, kwargs) self._inner = self.helper.create( self.name, *self._create_args, **kwargs)
python
def create(self, **kwargs): """ Create an instance of this resource definition. Only one instance may exist at any given time. """ if self.created: raise RuntimeError( '{} already created.'.format(self.__model_type__.__name__)) kwargs = self.merge_kwargs(self._create_kwargs, kwargs) self._inner = self.helper.create( self.name, *self._create_args, **kwargs)
[ "def", "create", "(", "self", ",", "**", "kwargs", ")", ":", "if", "self", ".", "created", ":", "raise", "RuntimeError", "(", "'{} already created.'", ".", "format", "(", "self", ".", "__model_type__", ".", "__name__", ")", ")", "kwargs", "=", "self", ".", "merge_kwargs", "(", "self", ".", "_create_kwargs", ",", "kwargs", ")", "self", ".", "_inner", "=", "self", ".", "helper", ".", "create", "(", "self", ".", "name", ",", "*", "self", ".", "_create_args", ",", "**", "kwargs", ")" ]
Create an instance of this resource definition. Only one instance may exist at any given time.
[ "Create", "an", "instance", "of", "this", "resource", "definition", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L54-L67
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
_DefinitionBase.remove
def remove(self, **kwargs): """ Remove an instance of this resource definition. """ self.helper.remove(self.inner(), **kwargs) self._inner = None
python
def remove(self, **kwargs): """ Remove an instance of this resource definition. """ self.helper.remove(self.inner(), **kwargs) self._inner = None
[ "def", "remove", "(", "self", ",", "**", "kwargs", ")", ":", "self", ".", "helper", ".", "remove", "(", "self", ".", "inner", "(", ")", ",", "**", "kwargs", ")", "self", ".", "_inner", "=", "None" ]
Remove an instance of this resource definition.
[ "Remove", "an", "instance", "of", "this", "resource", "definition", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L69-L74
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
_DefinitionBase.setup
def setup(self, helper=None, **create_kwargs): """ Setup this resource so that is ready to be used in a test. If the resource has already been created, this call does nothing. For most resources, this just involves creating the resource in Docker. :param helper: The resource helper to use, if one was not provided when this resource definition was created. :param **create_kwargs: Keyword arguments passed to :meth:`.create`. :returns: This definition instance. Useful for creating and setting up a resource in a single step:: volume = VolumeDefinition('volly').setup(helper=docker_helper) """ if self.created: return self.set_helper(helper) self.create(**create_kwargs) return self
python
def setup(self, helper=None, **create_kwargs): """ Setup this resource so that is ready to be used in a test. If the resource has already been created, this call does nothing. For most resources, this just involves creating the resource in Docker. :param helper: The resource helper to use, if one was not provided when this resource definition was created. :param **create_kwargs: Keyword arguments passed to :meth:`.create`. :returns: This definition instance. Useful for creating and setting up a resource in a single step:: volume = VolumeDefinition('volly').setup(helper=docker_helper) """ if self.created: return self.set_helper(helper) self.create(**create_kwargs) return self
[ "def", "setup", "(", "self", ",", "helper", "=", "None", ",", "**", "create_kwargs", ")", ":", "if", "self", ".", "created", ":", "return", "self", ".", "set_helper", "(", "helper", ")", "self", ".", "create", "(", "**", "create_kwargs", ")", "return", "self" ]
Setup this resource so that is ready to be used in a test. If the resource has already been created, this call does nothing. For most resources, this just involves creating the resource in Docker. :param helper: The resource helper to use, if one was not provided when this resource definition was created. :param **create_kwargs: Keyword arguments passed to :meth:`.create`. :returns: This definition instance. Useful for creating and setting up a resource in a single step:: volume = VolumeDefinition('volly').setup(helper=docker_helper)
[ "Setup", "this", "resource", "so", "that", "is", "ready", "to", "be", "used", "in", "a", "test", ".", "If", "the", "resource", "has", "already", "been", "created", "this", "call", "does", "nothing", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L76-L99
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
_DefinitionBase.as_fixture
def as_fixture(self, name=None): """ A decorator to inject this container into a function as a test fixture. """ if name is None: name = self.name def deco(f): @functools.wraps(f) def wrapper(*args, **kw): with self: kw[name] = self return f(*args, **kw) return wrapper return deco
python
def as_fixture(self, name=None): """ A decorator to inject this container into a function as a test fixture. """ if name is None: name = self.name def deco(f): @functools.wraps(f) def wrapper(*args, **kw): with self: kw[name] = self return f(*args, **kw) return wrapper return deco
[ "def", "as_fixture", "(", "self", ",", "name", "=", "None", ")", ":", "if", "name", "is", "None", ":", "name", "=", "self", ".", "name", "def", "deco", "(", "f", ")", ":", "@", "functools", ".", "wraps", "(", "f", ")", "def", "wrapper", "(", "*", "args", ",", "**", "kw", ")", ":", "with", "self", ":", "kw", "[", "name", "]", "=", "self", "return", "f", "(", "*", "args", ",", "**", "kw", ")", "return", "wrapper", "return", "deco" ]
A decorator to inject this container into a function as a test fixture.
[ "A", "decorator", "to", "inject", "this", "container", "into", "a", "function", "as", "a", "test", "fixture", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L147-L161
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.setup
def setup(self, helper=None, **run_kwargs): """ Creates the container, starts it, and waits for it to completely start. :param helper: The resource helper to use, if one was not provided when this container definition was created. :param **run_kwargs: Keyword arguments passed to :meth:`.run`. :returns: This container definition instance. Useful for creating and setting up a container in a single step:: con = ContainerDefinition('conny', 'nginx').setup(helper=dh) """ if self.created: return self.set_helper(helper) self.run(**run_kwargs) self.wait_for_start() return self
python
def setup(self, helper=None, **run_kwargs): """ Creates the container, starts it, and waits for it to completely start. :param helper: The resource helper to use, if one was not provided when this container definition was created. :param **run_kwargs: Keyword arguments passed to :meth:`.run`. :returns: This container definition instance. Useful for creating and setting up a container in a single step:: con = ContainerDefinition('conny', 'nginx').setup(helper=dh) """ if self.created: return self.set_helper(helper) self.run(**run_kwargs) self.wait_for_start() return self
[ "def", "setup", "(", "self", ",", "helper", "=", "None", ",", "**", "run_kwargs", ")", ":", "if", "self", ".", "created", ":", "return", "self", ".", "set_helper", "(", "helper", ")", "self", ".", "run", "(", "**", "run_kwargs", ")", "self", ".", "wait_for_start", "(", ")", "return", "self" ]
Creates the container, starts it, and waits for it to completely start. :param helper: The resource helper to use, if one was not provided when this container definition was created. :param **run_kwargs: Keyword arguments passed to :meth:`.run`. :returns: This container definition instance. Useful for creating and setting up a container in a single step:: con = ContainerDefinition('conny', 'nginx').setup(helper=dh)
[ "Creates", "the", "container", "starts", "it", "and", "waits", "for", "it", "to", "completely", "start", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L246-L267
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.teardown
def teardown(self): """ Stop and remove the container if it exists. """ while self._http_clients: self._http_clients.pop().close() if self.created: self.halt()
python
def teardown(self): """ Stop and remove the container if it exists. """ while self._http_clients: self._http_clients.pop().close() if self.created: self.halt()
[ "def", "teardown", "(", "self", ")", ":", "while", "self", ".", "_http_clients", ":", "self", ".", "_http_clients", ".", "pop", "(", ")", ".", "close", "(", ")", "if", "self", ".", "created", ":", "self", ".", "halt", "(", ")" ]
Stop and remove the container if it exists.
[ "Stop", "and", "remove", "the", "container", "if", "it", "exists", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L269-L276
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.status
def status(self): """ Get the container's current status from Docker. If the container does not exist (before creation and after removal), the status is ``None``. """ if not self.created: return None self.inner().reload() return self.inner().status
python
def status(self): """ Get the container's current status from Docker. If the container does not exist (before creation and after removal), the status is ``None``. """ if not self.created: return None self.inner().reload() return self.inner().status
[ "def", "status", "(", "self", ")", ":", "if", "not", "self", ".", "created", ":", "return", "None", "self", ".", "inner", "(", ")", ".", "reload", "(", ")", "return", "self", ".", "inner", "(", ")", ".", "status" ]
Get the container's current status from Docker. If the container does not exist (before creation and after removal), the status is ``None``.
[ "Get", "the", "container", "s", "current", "status", "from", "Docker", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L278-L288
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.stop
def stop(self, timeout=5): """ Stop the container. The container must have been created. :param timeout: Timeout in seconds to wait for the container to stop before sending a ``SIGKILL``. Default: 5 (half the Docker default) """ self.inner().stop(timeout=timeout) self.inner().reload()
python
def stop(self, timeout=5): """ Stop the container. The container must have been created. :param timeout: Timeout in seconds to wait for the container to stop before sending a ``SIGKILL``. Default: 5 (half the Docker default) """ self.inner().stop(timeout=timeout) self.inner().reload()
[ "def", "stop", "(", "self", ",", "timeout", "=", "5", ")", ":", "self", ".", "inner", "(", ")", ".", "stop", "(", "timeout", "=", "timeout", ")", "self", ".", "inner", "(", ")", ".", "reload", "(", ")" ]
Stop the container. The container must have been created. :param timeout: Timeout in seconds to wait for the container to stop before sending a ``SIGKILL``. Default: 5 (half the Docker default)
[ "Stop", "the", "container", ".", "The", "container", "must", "have", "been", "created", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L297-L306
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.run
def run(self, fetch_image=True, **kwargs): """ Create the container and start it. Similar to ``docker run``. :param fetch_image: Whether to try pull the image if it's not found. The behaviour here is similar to ``docker run`` and this parameter defaults to ``True``. :param **kwargs: Keyword arguments passed to :meth:`.create`. """ self.create(fetch_image=fetch_image, **kwargs) self.start()
python
def run(self, fetch_image=True, **kwargs): """ Create the container and start it. Similar to ``docker run``. :param fetch_image: Whether to try pull the image if it's not found. The behaviour here is similar to ``docker run`` and this parameter defaults to ``True``. :param **kwargs: Keyword arguments passed to :meth:`.create`. """ self.create(fetch_image=fetch_image, **kwargs) self.start()
[ "def", "run", "(", "self", ",", "fetch_image", "=", "True", ",", "**", "kwargs", ")", ":", "self", ".", "create", "(", "fetch_image", "=", "fetch_image", ",", "**", "kwargs", ")", "self", ".", "start", "(", ")" ]
Create the container and start it. Similar to ``docker run``. :param fetch_image: Whether to try pull the image if it's not found. The behaviour here is similar to ``docker run`` and this parameter defaults to ``True``. :param **kwargs: Keyword arguments passed to :meth:`.create`.
[ "Create", "the", "container", "and", "start", "it", ".", "Similar", "to", "docker", "run", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L308-L319
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.wait_for_start
def wait_for_start(self): """ Wait for the container to start. By default this will wait for the log lines matching the patterns passed in the ``wait_patterns`` parameter of the constructor using an UnorderedMatcher. For more advanced checks for container startup, this method should be overridden. """ if self.wait_matchers: matcher = UnorderedMatcher(*self.wait_matchers) self.wait_for_logs_matching(matcher, timeout=self.wait_timeout)
python
def wait_for_start(self): """ Wait for the container to start. By default this will wait for the log lines matching the patterns passed in the ``wait_patterns`` parameter of the constructor using an UnorderedMatcher. For more advanced checks for container startup, this method should be overridden. """ if self.wait_matchers: matcher = UnorderedMatcher(*self.wait_matchers) self.wait_for_logs_matching(matcher, timeout=self.wait_timeout)
[ "def", "wait_for_start", "(", "self", ")", ":", "if", "self", ".", "wait_matchers", ":", "matcher", "=", "UnorderedMatcher", "(", "*", "self", ".", "wait_matchers", ")", "self", ".", "wait_for_logs_matching", "(", "matcher", ",", "timeout", "=", "self", ".", "wait_timeout", ")" ]
Wait for the container to start. By default this will wait for the log lines matching the patterns passed in the ``wait_patterns`` parameter of the constructor using an UnorderedMatcher. For more advanced checks for container startup, this method should be overridden.
[ "Wait", "for", "the", "container", "to", "start", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L321-L332
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.get_logs
def get_logs(self, stdout=True, stderr=True, timestamps=False, tail='all', since=None): """ Get container logs. This method does not support streaming, use :meth:`stream_logs` for that. """ return self.inner().logs( stdout=stdout, stderr=stderr, timestamps=timestamps, tail=tail, since=since)
python
def get_logs(self, stdout=True, stderr=True, timestamps=False, tail='all', since=None): """ Get container logs. This method does not support streaming, use :meth:`stream_logs` for that. """ return self.inner().logs( stdout=stdout, stderr=stderr, timestamps=timestamps, tail=tail, since=since)
[ "def", "get_logs", "(", "self", ",", "stdout", "=", "True", ",", "stderr", "=", "True", ",", "timestamps", "=", "False", ",", "tail", "=", "'all'", ",", "since", "=", "None", ")", ":", "return", "self", ".", "inner", "(", ")", ".", "logs", "(", "stdout", "=", "stdout", ",", "stderr", "=", "stderr", ",", "timestamps", "=", "timestamps", ",", "tail", "=", "tail", ",", "since", "=", "since", ")" ]
Get container logs. This method does not support streaming, use :meth:`stream_logs` for that.
[ "Get", "container", "logs", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L400-L410
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.stream_logs
def stream_logs(self, stdout=True, stderr=True, tail='all', timeout=10.0): """ Stream container output. """ return stream_logs( self.inner(), stdout=stdout, stderr=stderr, tail=tail, timeout=timeout)
python
def stream_logs(self, stdout=True, stderr=True, tail='all', timeout=10.0): """ Stream container output. """ return stream_logs( self.inner(), stdout=stdout, stderr=stderr, tail=tail, timeout=timeout)
[ "def", "stream_logs", "(", "self", ",", "stdout", "=", "True", ",", "stderr", "=", "True", ",", "tail", "=", "'all'", ",", "timeout", "=", "10.0", ")", ":", "return", "stream_logs", "(", "self", ".", "inner", "(", ")", ",", "stdout", "=", "stdout", ",", "stderr", "=", "stderr", ",", "tail", "=", "tail", ",", "timeout", "=", "timeout", ")" ]
Stream container output.
[ "Stream", "container", "output", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L412-L418
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.wait_for_logs_matching
def wait_for_logs_matching(self, matcher, timeout=10, encoding='utf-8', **logs_kwargs): """ Wait for logs matching the given matcher. """ wait_for_logs_matching( self.inner(), matcher, timeout=timeout, encoding=encoding, **logs_kwargs)
python
def wait_for_logs_matching(self, matcher, timeout=10, encoding='utf-8', **logs_kwargs): """ Wait for logs matching the given matcher. """ wait_for_logs_matching( self.inner(), matcher, timeout=timeout, encoding=encoding, **logs_kwargs)
[ "def", "wait_for_logs_matching", "(", "self", ",", "matcher", ",", "timeout", "=", "10", ",", "encoding", "=", "'utf-8'", ",", "**", "logs_kwargs", ")", ":", "wait_for_logs_matching", "(", "self", ".", "inner", "(", ")", ",", "matcher", ",", "timeout", "=", "timeout", ",", "encoding", "=", "encoding", ",", "**", "logs_kwargs", ")" ]
Wait for logs matching the given matcher.
[ "Wait", "for", "logs", "matching", "the", "given", "matcher", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L420-L427
train
praekeltfoundation/seaworthy
seaworthy/definitions.py
ContainerDefinition.http_client
def http_client(self, port=None): """ Construct an HTTP client for this container. """ # Local import to avoid potential circularity. from seaworthy.client import ContainerHttpClient client = ContainerHttpClient.for_container(self, container_port=port) self._http_clients.append(client) return client
python
def http_client(self, port=None): """ Construct an HTTP client for this container. """ # Local import to avoid potential circularity. from seaworthy.client import ContainerHttpClient client = ContainerHttpClient.for_container(self, container_port=port) self._http_clients.append(client) return client
[ "def", "http_client", "(", "self", ",", "port", "=", "None", ")", ":", "from", "seaworthy", ".", "client", "import", "ContainerHttpClient", "client", "=", "ContainerHttpClient", ".", "for_container", "(", "self", ",", "container_port", "=", "port", ")", "self", ".", "_http_clients", ".", "append", "(", "client", ")", "return", "client" ]
Construct an HTTP client for this container.
[ "Construct", "an", "HTTP", "client", "for", "this", "container", "." ]
6f10a19b45d4ea1dc3bd0553cc4d0438696c079c
https://github.com/praekeltfoundation/seaworthy/blob/6f10a19b45d4ea1dc3bd0553cc4d0438696c079c/seaworthy/definitions.py#L429-L437
train
tuomas2/automate
src/automate/traits_fixes.py
_dispatch_change_event
def _dispatch_change_event(self, object, trait_name, old, new, handler): """ Prepare and dispatch a trait change event to a listener. """ # Extract the arguments needed from the handler. args = self.argument_transform(object, trait_name, old, new) # Send a description of the event to the change event tracer. if tnotifier._pre_change_event_tracer is not None: tnotifier._pre_change_event_tracer(object, trait_name, old, new, handler) # Dispatch the event to the listener. from automate.common import SystemNotReady try: self.dispatch(handler, *args) except SystemNotReady: pass except Exception as e: if tnotifier._post_change_event_tracer is not None: tnotifier._post_change_event_tracer(object, trait_name, old, new, handler, exception=e) # This call needs to be made inside the `except` block in case # the handler wants to re-raise the exception. tnotifier.handle_exception(object, trait_name, old, new) else: if tnotifier._post_change_event_tracer is not None: tnotifier._post_change_event_tracer(object, trait_name, old, new, handler, exception=None)
python
def _dispatch_change_event(self, object, trait_name, old, new, handler): """ Prepare and dispatch a trait change event to a listener. """ # Extract the arguments needed from the handler. args = self.argument_transform(object, trait_name, old, new) # Send a description of the event to the change event tracer. if tnotifier._pre_change_event_tracer is not None: tnotifier._pre_change_event_tracer(object, trait_name, old, new, handler) # Dispatch the event to the listener. from automate.common import SystemNotReady try: self.dispatch(handler, *args) except SystemNotReady: pass except Exception as e: if tnotifier._post_change_event_tracer is not None: tnotifier._post_change_event_tracer(object, trait_name, old, new, handler, exception=e) # This call needs to be made inside the `except` block in case # the handler wants to re-raise the exception. tnotifier.handle_exception(object, trait_name, old, new) else: if tnotifier._post_change_event_tracer is not None: tnotifier._post_change_event_tracer(object, trait_name, old, new, handler, exception=None)
[ "def", "_dispatch_change_event", "(", "self", ",", "object", ",", "trait_name", ",", "old", ",", "new", ",", "handler", ")", ":", "args", "=", "self", ".", "argument_transform", "(", "object", ",", "trait_name", ",", "old", ",", "new", ")", "if", "tnotifier", ".", "_pre_change_event_tracer", "is", "not", "None", ":", "tnotifier", ".", "_pre_change_event_tracer", "(", "object", ",", "trait_name", ",", "old", ",", "new", ",", "handler", ")", "from", "automate", ".", "common", "import", "SystemNotReady", "try", ":", "self", ".", "dispatch", "(", "handler", ",", "*", "args", ")", "except", "SystemNotReady", ":", "pass", "except", "Exception", "as", "e", ":", "if", "tnotifier", ".", "_post_change_event_tracer", "is", "not", "None", ":", "tnotifier", ".", "_post_change_event_tracer", "(", "object", ",", "trait_name", ",", "old", ",", "new", ",", "handler", ",", "exception", "=", "e", ")", "tnotifier", ".", "handle_exception", "(", "object", ",", "trait_name", ",", "old", ",", "new", ")", "else", ":", "if", "tnotifier", ".", "_post_change_event_tracer", "is", "not", "None", ":", "tnotifier", ".", "_post_change_event_tracer", "(", "object", ",", "trait_name", ",", "old", ",", "new", ",", "handler", ",", "exception", "=", "None", ")" ]
Prepare and dispatch a trait change event to a listener.
[ "Prepare", "and", "dispatch", "a", "trait", "change", "event", "to", "a", "listener", "." ]
d8a8cd03cd0da047e033a2d305f3f260f8c4e017
https://github.com/tuomas2/automate/blob/d8a8cd03cd0da047e033a2d305f3f260f8c4e017/src/automate/traits_fixes.py#L59-L85
train
cfobel/si-prefix
si_prefix/__init__.py
split
def split(value, precision=1): ''' Split `value` into value and "exponent-of-10", where "exponent-of-10" is a multiple of 3. This corresponds to SI prefixes. Returns tuple, where the second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Args ---- value : int, float Input value. precision : int Number of digits after decimal place to include. Returns ------- tuple The second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Examples -------- .. code-block:: python si_prefix.split(0.04781) -> (47.8, -3) si_prefix.split(4781.123) -> (4.8, 3) See :func:`si_format` for more examples. ''' negative = False digits = precision + 1 if value < 0.: value = -value negative = True elif value == 0.: return 0., 0 expof10 = int(math.log10(value)) if expof10 > 0: expof10 = (expof10 // 3) * 3 else: expof10 = (-expof10 + 3) // 3 * (-3) value *= 10 ** (-expof10) if value >= 1000.: value /= 1000.0 expof10 += 3 elif value >= 100.0: digits -= 2 elif value >= 10.0: digits -= 1 if negative: value *= -1 return value, int(expof10)
python
def split(value, precision=1): ''' Split `value` into value and "exponent-of-10", where "exponent-of-10" is a multiple of 3. This corresponds to SI prefixes. Returns tuple, where the second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Args ---- value : int, float Input value. precision : int Number of digits after decimal place to include. Returns ------- tuple The second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Examples -------- .. code-block:: python si_prefix.split(0.04781) -> (47.8, -3) si_prefix.split(4781.123) -> (4.8, 3) See :func:`si_format` for more examples. ''' negative = False digits = precision + 1 if value < 0.: value = -value negative = True elif value == 0.: return 0., 0 expof10 = int(math.log10(value)) if expof10 > 0: expof10 = (expof10 // 3) * 3 else: expof10 = (-expof10 + 3) // 3 * (-3) value *= 10 ** (-expof10) if value >= 1000.: value /= 1000.0 expof10 += 3 elif value >= 100.0: digits -= 2 elif value >= 10.0: digits -= 1 if negative: value *= -1 return value, int(expof10)
[ "def", "split", "(", "value", ",", "precision", "=", "1", ")", ":", "negative", "=", "False", "digits", "=", "precision", "+", "1", "if", "value", "<", "0.", ":", "value", "=", "-", "value", "negative", "=", "True", "elif", "value", "==", "0.", ":", "return", "0.", ",", "0", "expof10", "=", "int", "(", "math", ".", "log10", "(", "value", ")", ")", "if", "expof10", ">", "0", ":", "expof10", "=", "(", "expof10", "//", "3", ")", "*", "3", "else", ":", "expof10", "=", "(", "-", "expof10", "+", "3", ")", "//", "3", "*", "(", "-", "3", ")", "value", "*=", "10", "**", "(", "-", "expof10", ")", "if", "value", ">=", "1000.", ":", "value", "/=", "1000.0", "expof10", "+=", "3", "elif", "value", ">=", "100.0", ":", "digits", "-=", "2", "elif", "value", ">=", "10.0", ":", "digits", "-=", "1", "if", "negative", ":", "value", "*=", "-", "1", "return", "value", ",", "int", "(", "expof10", ")" ]
Split `value` into value and "exponent-of-10", where "exponent-of-10" is a multiple of 3. This corresponds to SI prefixes. Returns tuple, where the second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Args ---- value : int, float Input value. precision : int Number of digits after decimal place to include. Returns ------- tuple The second value is the "exponent-of-10" and the first value is `value` divided by the "exponent-of-10". Examples -------- .. code-block:: python si_prefix.split(0.04781) -> (47.8, -3) si_prefix.split(4781.123) -> (4.8, 3) See :func:`si_format` for more examples.
[ "Split", "value", "into", "value", "and", "exponent", "-", "of", "-", "10", "where", "exponent", "-", "of", "-", "10", "is", "a", "multiple", "of", "3", ".", "This", "corresponds", "to", "SI", "prefixes", "." ]
274fdf47f65d87d0b7a2e3c80f267db63d042c59
https://github.com/cfobel/si-prefix/blob/274fdf47f65d87d0b7a2e3c80f267db63d042c59/si_prefix/__init__.py#L47-L106
train