本文整理汇总了Python中scipy.empty_like函数的典型用法代码示例。如果您正苦于以下问题:Python empty_like函数的具体用法?Python empty_like怎么用?Python empty_like使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。
在下文中一共展示了empty_like函数的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Python代码示例。
示例1: density_2s
def density_2s(self, n1, n2):
"""Returns a reduced density matrix for a pair of sites.
Parameters
----------
n1 : int
The site number of the first site.
n2 : int
The site number of the second site (must be > n1).
"""
rho = sp.empty((self.q[n1] * self.q[n2], self.q[n1] * self.q[n2]), dtype=sp.complex128)
r_n2 = sp.empty_like(self.r[n2 - 1])
r_n1 = sp.empty_like(self.r[n1 - 1])
for s2 in xrange(self.q[n2]):
for t2 in xrange(self.q[n2]):
r_n2 = m.mmul(self.A[n2][t2], self.r[n2], m.H(self.A[n2][s2]))
r_n = r_n2
for n in reversed(xrange(n1 + 1, n2)):
r_n = self.eps_r(n, r_n)
for s1 in xrange(self.q[n1]):
for t1 in xrange(self.q[n1]):
r_n1 = m.mmul(self.A[n1][t1], r_n, m.H(self.A[n1][s1]))
tmp = m.mmul(self.l[n1 - 1], r_n1)
rho[s1 * self.q[n1] + s2, t1 * self.q[n1] + t2] = tmp.trace()
return rho
开发者ID:bcriger,项目名称:evoMPS,代码行数:28,代码来源:tdvp_gen.py
示例2: density_2s
def density_2s(self, n1, n2):
"""Returns a reduced density matrix for a pair of sites.
Currently only supports sites in the nonuniform window.
Parameters
----------
n1 : int
The site number of the first site.
n2 : int
The site number of the second site (must be > n1).
"""
rho = sp.empty((self.q[n1] * self.q[n2], self.q[n1] * self.q[n2]), dtype=sp.complex128)
r_n2 = sp.empty_like(self.r[n2 - 1])
r_n1 = sp.empty_like(self.r[n1 - 1])
ln1m1 = self.get_l(n1 - 1)
for s2 in xrange(self.q[n2]):
for t2 in xrange(self.q[n2]):
r_n2 = mm.mmul(self.A[n2][t2], self.r[n2], mm.H(self.A[n2][s2]))
r_n = r_n2
for n in reversed(xrange(n1 + 1, n2)):
r_n = tm.eps_r_noop(r_n, self.A[n], self.A[n])
for s1 in xrange(self.q[n1]):
for t1 in xrange(self.q[n1]):
r_n1 = mm.mmul(self.A[n1][t1], r_n, mm.H(self.A[n1][s1]))
tmp = mm.adot(ln1m1, r_n1)
rho[s1 * self.q[n1] + s2, t1 * self.q[n1] + t2] = tmp
return rho
开发者ID:hariseldon99,项目名称:evoMPS,代码行数:31,代码来源:mps_sandwich.py
示例3: calc_x
def calc_x(self, n, Vsh, sqrt_l, sqrt_r, sqrt_l_inv, sqrt_r_inv):
"""Calculate the parameter matrix x* giving the desired B.
This is equivalent to eqn. (49) of arXiv:1103.0936v2 [cond-mat.str-el] except
that, here, norm-preservation is not enforced, such that the optimal
parameter matrices x*_n (for the parametrization of B) are given by the
derivative w.r.t. x_n of <Phi[B, A]|Ĥ|Psi[A]>, rather than
<Phi[B, A]|Ĥ - H|Psi[A]> (with H = <Psi|Ĥ|Psi>).
An additional sum was added for the single-site hamiltonian.
Some multiplications have been pulled outside of the sums for efficiency.
Direct dependencies:
- A[n - 1], A[n], A[n + 1]
- r[n], r[n + 1], l[n - 2], l[n - 1]
- C[n], C[n - 1]
- K[n + 1]
- V[n]
"""
x = sp.zeros((self.D[n - 1], self.q[n] * self.D[n] - self.D[n - 1]), dtype=self.typ, order=self.odr)
x_part = sp.empty_like(x)
x_subpart = sp.empty_like(self.A[n][0])
x_subsubpart = sp.empty_like(self.A[n][0])
x_part.fill(0)
for s in xrange(self.q[n]):
x_subpart.fill(0)
if n < self.N:
x_subsubpart.fill(0)
for t in xrange(self.q[n + 1]):
x_subsubpart += m.mmul(self.C[n][s,t], self.r[n + 1], m.H(self.A[n + 1][t])) #~1st line
x_subsubpart += m.mmul(self.A[n][s], self.K[n + 1]) #~3rd line
x_subpart += m.mmul(x_subsubpart, sqrt_r_inv)
if not self.h_ext is None:
x_subsubpart.fill(0)
for t in xrange(self.q[n]): #Extra term to take care of h_ext..
x_subsubpart += self.h_ext(n, s, t) * self.A[n][t] #it may be more effecient to squeeze this into the nn term...
x_subpart += m.mmul(x_subsubpart, sqrt_r)
x_part += m.mmul(x_subpart, Vsh[s])
x += m.mmul(sqrt_l, x_part)
if n > 1:
x_part.fill(0)
for s in xrange(self.q[n]): #~2nd line
x_subsubpart.fill(0)
for t in xrange(self.q[n + 1]):
x_subsubpart += m.mmul(m.H(self.A[n - 1][t]), self.l[n - 2], self.C[n - 1][t, s])
x_part += m.mmul(x_subsubpart, sqrt_r, Vsh[s])
x += m.mmul(sqrt_l_inv, x_part)
return x
开发者ID:bcriger,项目名称:evoMPS,代码行数:58,代码来源:tdvp_gen.py
示例4: shifted_matrix_sub
def shifted_matrix_sub(data, sub, tau, pad_val=0.0):
"""Subtracts the multi-channeled vector (rows are channels) y from
the vector x with a certain offset. x and y can due to the offset be only
partly overlapping.
REM: from matlab
:type data: ndarray
:param data: data array to apply the subtractor to
:type sub: ndarray
:param sub: subtractor array
:type tau: int
:param tau: offset of :sub: w.r.t. start of :data:
:type pad_val: float
:param pad_val: value to use for the padding
Default=0.0
:return: ndarray - data minus sub at offset, len(data)
"""
ns_data, nc_data = data.shape
ns_sub, nc_sub = sub.shape
if nc_data != nc_sub:
raise ValueError('nc_data and nc_sub must agree!')
tau = int(tau)
data_sub = sp.empty_like(data)
data_sub[:] = pad_val
data_sub[max(0, tau):tau + ns_sub] = sub[max(0, -tau):ns_data - tau]
return data - data_sub
开发者ID:pmeier82,项目名称:BOTMpy,代码行数:28,代码来源:funcs_general.py
示例5: dispersion_relation_extraordinary
def dispersion_relation_extraordinary(kx, ky, k, nO, nE, c):
"""Dispersion relation for the extraordinary wave.
NOTE
See eq. 16 in Glytsis, "Three-dimensional (vector) rigorous
coupled-wave analysis of anisotropic grating diffraction",
JOSA A, 7(8), 1990 Always give positive real or negative
imaginary.
"""
if kx.shape != ky.shape or c.size != 3:
raise ValueError('kx and ky must have the same length and c must have 3 components')
kz = S.empty_like(kx)
for ii in xrange(0, kx.size):
alpha = nE**2 - nO**2
beta = kx[ii]/k * c[0] + ky[ii]/k * c[1]
# coeffs
C = S.array([nO**2 + c[2]**2 * alpha, \
2. * c[2] * beta * alpha, \
nO**2 * (kx[ii]**2 + ky[ii]**2) / k**2 + alpha * beta**2 - nO**2 * nE**2])
# two solutions of type +x or -x, purely real or purely imag
tmp_kz = k * S.roots(C)
# get the negative imaginary part or the positive real one
if S.any(S.isreal(tmp_kz)):
kz[ii] = S.absolute(tmp_kz[0])
else:
kz[ii] = -1j * S.absolute(tmp_kz[0])
return kz
开发者ID:LeiDai,项目名称:EMpy,代码行数:35,代码来源:RCWA.py
示例6: predict_proba
def predict_proba(self, X):
"""
Predict the membership probabilities for the data samples
in X using trained model.
Parameters
----------
X : array-like, shape (n_samples, n_features)
List of n_features-dimensional data points. Each row
corresponds to a single data point.
Returns
-------
proba : array, shape (n_samples, n_clusters)
"""
X = check_array(X, copy=False, order='C', dtype=sp.float64)
K = self.score_samples(X)
T = sp.empty_like(K)
# Compute the Loglikelhood
K *= (0.5)
# Compute the posterior
with sp.errstate(over='ignore'):
for c in xrange(self.C):
T[:, c] = 1 / sp.exp(K-K[:, c][:, sp.newaxis]).sum(axis=1)
return T
开发者ID:mfauvel,项目名称:HDDA,代码行数:28,代码来源:hdda.py
示例7: scale
def scale(self,x,M=None,m=None): # TODO: DO IN PLACE SCALING
"""[email protected] Function that standardize the data
Input:
x: the data
M: the Max vector
m: the Min vector
Output:
x: the standardize data
M: the Max vector
m: the Min vector
"""
[n,d]=x.shape
if not sp.issubdtype(x.dtype,float):
x=x.astype('float')
# Initialization of the output
xs = sp.empty_like(x)
# get the parameters of the scaling
if M is None:
M,m = sp.amax(x,axis=0),sp.amin(x,axis=0)
den = M-m
for i in range(d):
if den[i] != 0:
xs[:,i] = 2*(x[:,i]-m[i])/den[i]-1
else:
xs[:,i]=x[:,i]
return xs
开发者ID:lennepkade,项目名称:HistoricalMap,代码行数:31,代码来源:function_historical_map.py
示例8: calc_B
def calc_B(self, n, set_eta=True):
"""Generates the B[n] tangent vector corresponding to physical evolution of the state.
In other words, this returns B[n][x*] (equiv. eqn. (47) of
arXiv:1103.0936v2 [cond-mat.str-el])
with x* the parameter matrices satisfying the Euler-Lagrange equations
as closely as possible.
In the case of Bc, use the general Bc generated in calc_B_centre().
"""
if n == self.N_centre:
B, eta_sq_c = self.calc_B_centre()
if set_eta:
self.eta_sq[self.N_centre] = eta_sq_c
else:
l_sqrt, r_sqrt, l_sqrt_inv, r_sqrt_inv = self.calc_l_r_roots(n)
if n > self.N_centre:
Vsh = tm.calc_Vsh(self.A[n], r_sqrt, sanity_checks=self.sanity_checks)
x = self.calc_x(n, Vsh, l_sqrt, r_sqrt, l_sqrt_inv, r_sqrt_inv, right=True)
B = sp.empty_like(self.A[n])
for s in range(self.q[n]):
B[s] = mm.mmul(l_sqrt_inv, x, mm.H(Vsh[s]), r_sqrt_inv)
if self.sanity_checks:
M = tm.eps_r_noop(self.r[n], B, self.A[n])
if not sp.allclose(M, 0):
print("Sanity Fail in calc_B!: B_%u does not satisfy GFC!" % n)
else:
Vsh = tm.calc_Vsh_l(self.A[n], l_sqrt, sanity_checks=self.sanity_checks)
x = self.calc_x(n, Vsh, l_sqrt, r_sqrt, l_sqrt_inv, r_sqrt_inv, right=False)
B = sp.empty_like(self.A[n])
for s in range(self.q[n]):
B[s] = mm.mmul(l_sqrt_inv, mm.H(Vsh[s]), x, r_sqrt_inv)
if self.sanity_checks:
M = tm.eps_l_noop(self.l[n - 1], B, self.A[n])
if not sp.allclose(M, 0):
print("Sanity Fail in calc_B!: B_%u does not satisfy GFC!" % n)
if set_eta:
self.eta_sq[n] = mm.adot(x, x)
return B
开发者ID:amilsted,项目名称:evoMPS,代码行数:47,代码来源:tdvp_sandwich.py
示例9: calc_B1
def calc_B1(self):
"""Calculate the optimal B1 given right gauge-fixing on B2..N and
no gauge-fixing on B1.
We use the non-norm-preserving K's, since the norm-preservation
is not needed elsewhere. It is cleaner to subtract the relevant
norm-changing terms from the K's here than to generate all K's
with norm-preservation.
"""
B1 = sp.empty_like(self.A[1])
try:
r1_i = self.r[1].inv()
except AttributeError:
r1_i = mm.invmh(self.r[1])
try:
l0_i = self.l[0].inv()
except AttributeError:
l0_i = mm.invmh(self.l[0])
A0 = self.A[0]
A1 = self.A[1]
A2 = self.A[2]
r1 = self.r[1]
r2 = self.r[2]
l0 = self.l[0]
KLh = mm.H(self.u_gnd_l.K_left - l0 * mm.adot(self.u_gnd_l.K_left, self.r[0]))
K2 = self.K[2] - r1 * mm.adot(self.l[1], self.K[2])
C1 = self.C[1] - self.h_expect[1] * self.AA1
C0 = self.C[0] - self.h_expect[0] * self.AA0
for s in xrange(self.q[1]):
try:
B1[s] = A1[s].dot(r1_i.dot_left(K2))
except AttributeError:
B1[s] = A1[s].dot(K2.dot(r1_i))
for t in xrange(self.q[2]):
try:
B1[s] += C1[s, t].dot(r2.dot(r1_i.dot_left(mm.H(A2[t]))))
except AttributeError:
B1[s] += C1[s, t].dot(r2.dot(mm.H(A2[t]).dot(r1_i)))
B1sbit = KLh.dot(A1[s])
for t in xrange(self.q[0]):
B1sbit += mm.H(A0[t]).dot(l0.dot(C0[t,s]))
B1[s] += l0_i.dot(B1sbit)
rb = sp.zeros_like(self.r[0])
for s in xrange(self.q[1]):
rb += B1[s].dot(r1.dot(mm.H(B1[s])))
eta = sp.sqrt(mm.adot(l0, rb))
return B1, eta
开发者ID:bcriger,项目名称:evoMPS,代码行数:59,代码来源:tdvp_sandwich.py
示例10: calculate_slow_phi_0s
def calculate_slow_phi_0s(phi_0s, p_values):
slow_phi_0s = scipy.empty_like(phi_0s)
for i, phi_0 in enumerate(phi_0s):
phi_0_unfolded = unfold(phi_0)
x = arange(len(phi_0_unfolded))
model = polyfit(x, phi_0_unfolded, 3, w=p_values[i])
slow_phi_0s[i] = polyval(model, x)
return slow_phi_0s
开发者ID:martina88esposito,项目名称:tomohowk,代码行数:8,代码来源:correct_slow_phi_0.py
示例11: calc_C
def calc_C(self, n_low=-1, n_high=-1):
"""Generates the C matrices used to calculate the K's and ultimately the B's
These are to be used on one side of the super-operator when applying the
nearest-neighbour Hamiltonian, similarly to C in eqn. (44) of
arXiv:1103.0936v2 [cond-mat.str-el], except being for the non-norm-preserving case.
Makes use only of the nearest-neighbour hamiltonian, and of the A's.
C[n] depends on A[n] and A[n + 1].
This calculation can be significantly faster if a matrix form for h_nn
is available. See gen_h_matrix().
"""
if self.h_nn is None:
return 0
if n_low < 1:
n_low = 0
if n_high < 1:
n_high = self.N + 1
if self.h_nn_mat is None:
for n in xrange(n_low, n_high):
self.C[n].fill(0)
for u in xrange(self.q[n]):
for v in xrange(self.q[n + 1]):
AA = mm.mmul(self.A[n][u], self.A[n + 1][v]) #only do this once for each
for s in xrange(self.q[n]):
for t in xrange(self.q[n + 1]):
h_nn_stuv = self.h_nn(n, s, t, u, v)
if h_nn_stuv != 0:
self.C[n][s, t] += h_nn_stuv * AA
else:
dot = sp.dot
for n in xrange(n_low, n_high):
An = self.A[n]
Anp1 = self.A[n + 1]
AA = sp.empty_like(self.C[n])
for u in xrange(self.q[n]):
for v in xrange(self.q[n + 1]):
AA[u, v] = dot(An[u], Anp1[v])
if n == 0: #FIXME: Temp. hack
self.AA0 = AA
elif n == 1:
self.AA1 = AA
res = sp.tensordot(AA, self.h_nn_mat[n], ((0, 1), (2, 3)))
res = sp.rollaxis(res, 3)
res = sp.rollaxis(res, 3)
self.C[n][:] = res
开发者ID:bcriger,项目名称:evoMPS,代码行数:55,代码来源:tdvp_sandwich.py
示例12: oppwalker_convert
def oppwalker_convert(arr):
#assert(arr.min()>=0 and arr.max()<=1)
out = sp.empty_like(arr)
# red-green
out[:,:,0] = arr[:,:,0] - arr[:,:,1]
# blue-yellow
out[:,:,1] = arr[:,:,2] - arr[:,:,[0,1]].min(2)
# intensity
out[:,:,2] = arr.max(2)
return out
开发者ID:aparicio,项目名称:v1like,代码行数:12,代码来源:colorconv.py
示例13: BiotLineIntegral
def BiotLineIntegral(vect_arr, r_p, current=1.0):
"""
Calculates the magnetic flux density of from a list of points
(vect_arr) that represent a discretization of a conductor.
The magnetic flux density is calculated for r_p positions.
Parameters
----------
vect_arr: array([[x_0,y_0,z_0], ... , [x_n,y_n,z_n]])
r_p: array([[x_0,y_0,z_0], ... , [x_n,y_n,z_n]])
All coordinates are in mm
Returns
----------
Magnetric flux density Bfield, same as r_p
"""
bfield = scipy.empty_like(r_p, scipy.float64)
code = """
#include <iostream>
#include <math.h>
#include <assert.h>
for(int ir_p = 0; ir_p < size_r_p; ir_p++) {
vec3 wire_pre = { vect_arr(0,0), vect_arr(0,1), vect_arr(0,2) };
vec3 bfield_vec = { 0.0, 0.0, 0.0 };
vec3 vec3r_p = { r_p(ir_p, 0), r_p(ir_p, 1), r_p(ir_p, 2) };
for(int i_v = 1; i_v < size_vect_arr; i_v++) {
vec3 vec3_arr = { vect_arr(i_v,0), vect_arr(i_v,1), vect_arr(i_v,2) };
vec3 dl = vec3_diff( vec3_arr, wire_pre);
vec3 rs = vec3_arr;
vec3 r = vec3_diff( vec3r_p, rs );
double r_length = vec3_abs(r);
bfield_vec = vec3_add( vec3_scale( vec3_cross( dl, r), 1.0 / pow( r_length, 3)), bfield_vec );
wire_pre = vec3_arr;
}
bfield(ir_p, 0) = bfield_vec.x ;
bfield(ir_p, 1) = bfield_vec.y ;
bfield(ir_p, 2) = bfield_vec.z ;
}
return_val = 1;
"""
size_r_p = r_p[:, 0].size
size_vect_arr = vect_arr[:, 0].size
os.path.realpath(__file__)
support_code = open( os.path.dirname(__file__) + "/biot_blitz_support.cpp" )
scipy.weave.inline(code,
["r_p", "size_r_p", "bfield", "vect_arr", "size_vect_arr"],
type_converters=converters.blitz,
support_code=support_code.read(),
compiler='gcc' )
return bfield * mu_0 * 1000.0 * 1.0/(4.0 * pi)
开发者ID:marmei,项目名称:FreeCADBiotSavart,代码行数:52,代码来源:BiotLineIntegral.py
示例14: oppsande_convert
def oppsande_convert(arr):
#assert(arr.min()>=0 and arr.max()<=1)
r = arr[:,:,0]
g = arr[:,:,1]
b = arr[:,:,2]
out = sp.empty_like(arr)
out[:,:,0] = (r-g) / sp.sqrt(2.)
out[:,:,1] = (r+g-2.*b) / sp.sqrt(6.)
out[:,:,2] = (r+g+b) / sp.sqrt(3.)
return out
开发者ID:aparicio,项目名称:v1like,代码行数:13,代码来源:colorconv.py
示例15: invE_convert
def invE_convert(arr):
#assert(arr.min()>=0 and arr.max()<=1)
red = arr[:,:,0]
green = arr[:,:,1]
blue = arr[:,:,2]
out = sp.empty_like(arr)
out[:,:,0] = (red + green + blue) / 3.
out[:,:,1] = (red + green - 2.*blue) / 4.
out[:,:,2] = (red - 2.*green + blue) / 4.
return out
开发者ID:aparicio,项目名称:v1like,代码行数:14,代码来源:colorconv.py
示例16: van_rossum_dist
def van_rossum_dist(trains, tau=1.0 * pq.s, kernel=None, sort=True):
""" Calculates the van Rossum distance.
It is defined as Euclidean distance of the spike trains convolved with a
causal decaying exponential smoothing filter. A detailed description can be
found in *Rossum, M. C. W. (2001). A novel spike distance. Neural
Computation, 13(4), 751-763.* This implementation is normalized to yield
a distance of 1.0 for the distance between an empty spike train and a spike
train with a single spike. Divide the result by sqrt(2.0) to get the
normalization used in the cited paper.
Given :math:`N` spike trains with :math:`n` spikes on average the run-time
complexity of this function is :math:`O(N^2 n^2)`. An implementation in
:math:`O(N^2 n)` would be possible but has a high constant factor rendering
it slower in practical cases.
:param sequence trains: Sequence of :class:`neo.core.SpikeTrain` objects of
which the van Rossum distance will be calculated pairwise.
:param tau: Decay rate of the exponential function as time scalar. Controls
for which time scale the metric will be sensitive. This parameter will
be ignored if `kernel` is not `None`. May also be :const:`scipy.inf`
which will lead to only measuring differences in spike count.
:type tau: Quantity scalar
:param kernel: Kernel to use in the calculation of the distance. This is not
the smoothing filter, but its autocorrelation. If `kernel` is `None`, an
unnormalized Laplacian kernel with a size of `tau` will be used.
:type kernel: :class:`.signal_processing.Kernel`
:param bool sort: Spike trains with sorted spike times might be needed for
the calculation. You can set `sort` to `False` if you know that your
spike trains are already sorted to decrease calculation time.
:returns: Matrix containing the van Rossum distances for all pairs of spike
trains.
:rtype: 2-D array
"""
if kernel is None:
if tau == sp.inf:
spike_counts = [st.size for st in trains]
return (spike_counts - sp.atleast_2d(spike_counts).T) ** 2
kernel = sigproc.LaplacianKernel(tau, normalize=False)
k_dist = kernel.summed_dist_matrix(
[st.view(type=pq.Quantity) for st in trains], not sort)
vr_dist = sp.empty_like(k_dist)
for i, j in sp.ndindex(*k_dist.shape):
vr_dist[i, j] = (
k_dist[i, i] + k_dist[j, j] - k_dist[i, j] - k_dist[j, i])
return sp.sqrt(vr_dist)
开发者ID:NeuroArchive,项目名称:spykeutils,代码行数:48,代码来源:spike_train_metrics.py
示例17: density_1s
def density_1s(self, n):
"""Returns a reduced density matrix for a single site.
Parameters
----------
n1 : int
The site number.
"""
rho = sp.empty((self.q[n], self.q[n]), dtype=sp.complex128)
r_n = self.r[n]
r_nm1 = sp.empty_like(self.r[n - 1])
for s in xrange(self.q[n]):
for t in xrange(self.q[n]):
r_nm1 = m.mmul(self.A[n][t], r_n, m.H(self.A[n][s]))
rho[s, t] = m.mmul(self.l[n - 1], r_nm1).trace()
return rho
开发者ID:bcriger,项目名称:evoMPS,代码行数:17,代码来源:tdvp_gen.py
示例18: _calc_B_r_diss
def _calc_B_r_diss(self, op, K, C, n, set_eta=True):
if self.q[n] * self.D[n] - self.D[n - 1] > 0:
l_sqrt, l_sqrt_inv, r_sqrt, r_sqrt_inv = tm.calc_l_r_roots(self.l[n - 1],
self.r[n],
sanity_checks=self.sanity_checks,
sc_data=("site", n))
Vsh = tm.calc_Vsh(self.A[n], r_sqrt, sanity_checks=self.sanity_checks)
x = self.calc_x(n, Vsh, l_sqrt, r_sqrt, l_sqrt_inv, r_sqrt_inv)
if set_eta:
self.eta[n] = sp.sqrt(mm.adot(x, x))
B = sp.empty_like(self.A[n])
for s in xrange(self.q[n]):
B[s] = mm.mmul(l_sqrt_inv, x, mm.H(Vsh[s]), r_sqrt_inv)
return B
else:
return None
开发者ID:ftranschel,项目名称:evoMPS,代码行数:17,代码来源:fermi-hubbard.py
示例19: rg2_convert
def rg2_convert(arr):
#assert(arr.min()>=0 and arr.max()<=1)
out = sp.empty_like(arr[:,:,[0,1]])
red = arr[:,:,0]
green = arr[:,:,1]
blue = arr[:,:,2]
intensity = arr.mean(2)
lowi = intensity < 0.1*intensity.max()
arr[lowi] = 0
denom = arr.sum(2)
denom[denom==0] = 1
out[:,:,0] = red / denom
out[:,:,1] = green / denom
return out
开发者ID:aparicio,项目名称:v1like,代码行数:19,代码来源:colorconv.py
示例20: chrom_convert
def chrom_convert(arr):
#assert(arr.min()>=0 and arr.max()<=1)
opp = opp_convert(arr)
out = sp.empty_like(opp[:,:,[0,1]])
rg = opp[:,:,0]
by = opp[:,:,1]
intensity = opp[:,:,2]
lowi = intensity < 0.1*intensity.max()
rg[lowi] = 0
by[lowi] = 0
denom = intensity
denom[denom==0] = 1
out[:,:,0] = rg / denom
out[:,:,1] = by / denom
return out
开发者ID:aparicio,项目名称:v1like,代码行数:20,代码来源:colorconv.py
注:本文中的scipy.empty_like函数示例由纯净天空整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论