p = 10 mean = np.zeros(p) sigma = np.zeros((p, p)) for i in range(p): for j in range(p): sigma[i, j] = np.exp(-abs(i-j)) mydata = mvn(mean, sigma, size=1000) #生成数据 beta = mvn(0.2np.ones(p),0.01sigma,size=6) #生成后60%数据的β for p2 in range(2,4): mydata[:, p2] = np.where(mydata[:, p2] > 0, 1, 0) #将第二个20%的数据转化为二进制 for p3 in range(5,p): mydata[:, p3] = np.where(mydata[:, p3] > 0.5, 2, np.where(mydata[:, p3] <= -0.5, 0, 1)) U = uniform.rvs(size=1000) U = 1 - U U = np.log(U) J = np.diag(U) # 生成1000个随机数满足(0,1)的均匀分布 for p40 in range(0,4): A11=0.2mydata[:, :p40].sum(axis=1) for p60 in range(4,p): A12=0.2mydata[:, :p60].sum(axis=1) A1=A11+A12 #情景1 B = 10*((0.01)**10) * np.exp(A1) E = np.eye(1000) T = np.zeros((1000, 1)) for m in range(1000): T[m, 0] = J[m, m]/ B T = np.sqrt(T).T 删失时间 Ci = np.random.exponential(scale=7, size=1000).reshape(-1, 1) time = np.zeros((1000, 1)) for m in range(1000): if T[0, m] > Ci[m, 0]: time[m, 0] = Ci[m, 0] else: time[m, 0] = T E = np.zeros((1000, 1)) for m in range(1000): if T[0, m] > Ci[m, 0]: E[m, 0] = 0 else: E[m, 0] = 1 X=mydata Y=time E=E 为什么得出来的T是nan

时间: 2024-04-12 14:29:16 浏览: 20
在给定的代码段中,变量`B`的计算涉及到除以`A1`,而`A1`是一个长度为1000的数组。如果在`A1`中存在0元素,那么除以0将会导致结果为无穷大(inf),进而导致计算出的`T`变量中存在NaN值。 为了解决这个问题,你可以在计算`B`之前,先检查`A1`中是否存在0元素,并将其替换为一个较小的非零值。例如,你可以使用以下代码来替代计算`B`的那一行: ```python A1 = A11 + A12 # 情景1 A1 = np.where(A1 == 0, 1e-10, A1) # 将A1中的0替换为较小的非零值 B = 10 * ((0.01) ** 10) * np.exp(A1) ``` 这样做可以避免除以0导致的NaN值问题。请注意,我在代码中使用了一个较小的非零值(1e-10),你可以根据具体情况调整这个值。
相关问题

import numpy as np import random from scipy import stats import pandas as pd import matplotlib.pyplot as plt np.random.seed(1) a=[] for p in range(1,11): k=8 n=100 Sigma = [[1,0.6+0.04p],[0.6+0.04p,1]] res1 = [] for i in range(1,1001): data=np.random.multivariate_normal(np.zeros(2), Sigma, n) X_data=data[:,0] Y_data=data[:,1] Sx=sorted(X_data) Sy=sorted(Y_data) inter_x=np.arange(min(X_data),max(X_data)+(max(X_data)-min(X_data))/k, (max(X_data)-min(X_data))/k) inter_y=np.arange(min(Y_data),max(Y_data)+(max(Y_data)-min(Y_data))/k, (max(Y_data)-min(Y_data))/k) left_inter_x=np.dot(np.ones((n,1)),inter_x[0:k].reshape(1,k)) right_inter_x=np.dot(np.ones((n,1)),inter_x[1:(k+1)].reshape(1,k)) left_inter_y=np.dot(np.ones((n,1)),inter_y[0:k].reshape(1,k)) right_inter_y=np.dot(np.ones((n,1)),inter_y[1:(k+1)].reshape(1,k)) Data1=np.dot(X_data.reshape(n,1), np.ones((1,k))) Data2=np.dot(Y_data.reshape(n,1), np.ones((1,k))) frequx=(left_inter_x<=Data1)(Data1<right_inter_x) frequy=(left_inter_y<=Data2)(Data2<right_inter_y) frequxy = np.dot(frequx.astype(int).T,frequy.astype(int)) pi=np.sum(frequxy,axis=0)/n pj=np.sum(frequxy,axis=1)/n pij=np.dot(pi.reshape(k,1),pj.reshape(1,k)) A=(frequxy-npij)**2/(npij) A[np.isnan(A)]=0 A[np.isinf(A)]=0 stat1=np.sum(A) res1.append(int(stat1>stats.chi2.ppf(0.95,(k-1)**2))) a[p]=np.mean(res1) plt.plot(a)有哪些错误

该代码存在以下错误: 1. 缺少函数声明和主程序入口。 2. 行首缺少空格,不符合 Python 的 PEP8 代码规范。 3. 第一行导入 numpy 库和 random 库,但是代码中没有使用 random 库,可以删除。 4. 缺少对 scipy 库和 pandas 库的导入。 5. 第5行代码中的 Sigma 变量没有正确的缩进,应该缩进4个空格。 6. 第6行代码中的 res1 变量没有正确的缩进,应该缩进4个空格。 7. 第7行代码中应该在 import 后面加上空格。 8. 第9行代码中的 np.random.seed(1) 应该缩进4个空格。 9. 第11行代码中的 a=[] 应该缩进4个空格。 10. 第13行代码中的 p 变量没有正确的缩进,应该缩进4个空格。 11. 第14行代码中的 n 变量没有正确的缩进,应该缩进4个空格。 12. 第15行代码中的 Sigma 变量应该缩进4个空格。 13. 第16-30行代码是一个 for 循环,缩进不正确,应该缩进4个空格。 14. 第17-29行代码中的变量应该缩进8个空格。 15. 第18-20行代码中的 X_data, Y_data, Sx, Sy, inter_x, inter_y 等变量没有正确的缩进,应该缩进8个空格。 16. 第21-24行代码中的 left_inter_x, right_inter_x, left_inter_y, right_inter_y 等变量没有正确的缩进,应该缩进8个空格。 17. 第26行代码中的 Data1, Data2 变量没有正确的缩进,应该缩进8个空格。 18. 第27-28行代码中的 frequx, frequy 变量没有正确的缩进,应该缩进8个空格。 19. 第29行代码中的 frequxy 变量没有正确的缩进,应该缩进8个空格。 20. 第30行代码中的 pi, pj, pij, A 变量没有正确的缩进,应该缩进8个空格。 21. 第32行代码中的 np.isnan(A) 和 np.isinf(A) 应该缩进8个空格。 22. 第34行代码中的 a[p]=np.mean(res1) 应该缩进4个空格。 23. 最后一行代码中的 plt.plot(a) 应该缩进4个空格。 建议修改后的代码如下所示: ```python import numpy as np from scipy import stats import pandas as pd import matplotlib.pyplot as plt def main(): np.random.seed(1) a = [0] * 11 for p in range(1, 11): k = 8 n = 100 Sigma = [[1, 0.6 + 0.04 * p], [0.6 + 0.04 * p, 1]] res1 = [] for i in range(1, 1001): data = np.random.multivariate_normal(np.zeros(2), Sigma, n) X_data = data[:, 0] Y_data = data[:, 1] Sx = sorted(X_data) Sy = sorted(Y_data) inter_x = np.arange(min(X_data), max(X_data) + (max(X_data) - min(X_data)) / k, (max(X_data) - min( X_data)) / k) inter_y = np.arange(min(Y_data), max(Y_data) + (max(Y_data) - min(Y_data)) / k, (max(Y_data) - min( Y_data)) / k) left_inter_x = np.dot(np.ones((n, 1)), inter_x[0:k].reshape(1, k)) right_inter_x = np.dot(np.ones((n, 1)), inter_x[1:(k + 1)].reshape(1, k)) left_inter_y = np.dot(np.ones((n, 1)), inter_y[0:k].reshape(1, k)) right_inter_y = np.dot(np.ones((n, 1)), inter_y[1:(k + 1)].reshape(1, k)) Data1 = np.dot(X_data.reshape(n, 1), np.ones((1, k))) Data2 = np.dot(Y_data.reshape(n, 1), np.ones((1, k))) frequx = (left_inter_x <= Data1) * (Data1 < right_inter_x) frequy = (left_inter_y <= Data2) * (Data2 < right_inter_y) frequxy = np.dot(frequx.astype(int).T, frequy.astype(int)) pi = np.sum(frequxy, axis=0) / n pj = np.sum(frequxy, axis=1) / n pij = np.dot(pi.reshape(k, 1), pj.reshape(1, k)) npij = n * pij A = (frequxy - npij) ** 2 / (npij) A[np.isnan(A)] = 0 A[np.isinf(A)] = 0 stat1 = np.sum(A) res1.append(int(stat1 > stats.chi2.ppf(0.95, (k - 1) ** 2))) a[p] = np.mean(res1) plt.plot(a) if __name__ == '__main__': main() ```

显示代码中y_rec的函数表达式:import numpy as np import matplotlib.pyplot as plt def gen_data(x1, x2): y_sample = np.sin(np.pi * x1 / 2) + np.cos(np.pi * x1 / 3) y_all = np.sin(np.pi * x2 / 2) + np.cos(np.pi * x2 / 3) return y_sample, y_all def kernel_interpolation(y_sample, x1, sig): gaussian_kernel = lambda x, c, h: np.exp(-(x - x[c]) ** 2 / (2 * (h ** 2))) num = len(y_sample) w = np.zeros(num) int_matrix = np.asmatrix(np.zeros((num, num))) for i in range(num): int_matrix[i, :] = gaussian_kernel(x1, i, sig) w = int_matrix.I * np.asmatrix(y_sample).T return w def kernel_interpolation_rec(w, x1, x2, sig): gkernel = lambda x, xc, h: np.exp(-(x - xc) ** 2 / (2 * (h ** 2))) num = len(x2) y_rec = np.zeros(num) for i in range(num): for k in range(len(w)): y_rec[i] = y_rec[i] + w[k] * gkernel(x2[i], x1[k], sig) return y_rec if name == 'main': snum =4 # control point数量 ratio =50 # 总数据点数量:snum*ratio sig = 2 # 核函数宽度 xs = -14 xe = 14 #x1 = np.linspace(xs, xe,snum) x1 = np.array([9, 9.1, 13 ]) x2 = np.linspace(xs, xe, (snum - 1) * ratio + 1) y_sample, y_all = gen_data(x1, x2) plt.figure(1) w = kernel_interpolation(y_sample, x1, sig) y_rec = kernel_interpolation_rec(w, x1, x2, sig) plt.plot(x2, y_rec, 'k') plt.plot(x2, y_all, 'r:') plt.ylabel('y') plt.xlabel('x') for i in range(len(x1)): plt.plot(x1[i], y_sample[i], 'go', markerfacecolor='none') # 计算均方根误差 rmse = np.sqrt(np.mean((y_rec - y_all) ** 2)) # 输出均方根误差值 print("均方根误差为:", rmse) plt.legend(labels=['reconstruction', 'original', 'control point'], loc='lower left') plt.title('kernel interpolation:$y=sin(\pi x/2)+cos(\pi x/3)$') plt.show()

代码中 y_rec 的函数表达式为: ```python def kernel_interpolation_rec(w, x1, x2, sig): gkernel = lambda x, xc, h: np.exp(-(x - xc) ** 2 / (2 * (h ** 2))) num = len(x2) y_rec = np.zeros(num) for i in range(num): for k in range(len(w)): y_rec[i] = y_rec[i] + w[k] * gkernel(x2[i], x1[k], sig) return y_rec ``` 其中,$w$ 是控制点的权重,$x_1$ 是控制点的横坐标,$x_2$ 是所有数据点的横坐标,$\sigma$ 是高斯核函数的宽度,$gkernel$ 是高斯核函数。在函数中,我们遍历所有数据点的横坐标 $x_2$,对于每个数据点,我们遍历所有控制点的权重 $w$,然后将每个控制点对当前数据点的贡献累加起来,得到在当前数据点处通过核插值重构出来的纵坐标 $y_{rec}[i]$。

相关推荐

import torch import torch.nn as nn import torch.nn.functional as F import torchvision.transforms as transforms import skimage.segmentation as seg import numpy as np # 超参数 from PIL import Image num_superpixels = 1000 compactness = 10 sigma = 1 # 定义模型 class SuperpixelSegmentation(nn.Module): def init(self): super(SuperpixelSegmentation, self).init() self.convs = nn.Sequential( nn.Conv2d(3, 16, kernel_size=3, stride=1, padding=1), nn.ReLU(inplace=True), nn.Conv2d(16, 32, kernel_size=3, stride=1, padding=1), nn.ReLU(inplace=True), nn.Conv2d(32, 64, kernel_size=3, stride=1, padding=1), nn.ReLU(inplace=True), nn.Conv2d(64, num_superpixels, kernel_size=1, stride=1) ) def forward(self, x): x = self.convs(x) return x # 加载图像 imgA = Image.open('1.png').convert('RGB') imgB = Image.open('2.jpg').convert('RGB') # 超像素分割 imgA_np = np.array(imgA) segments = seg.slic(imgA_np, n_segments=num_superpixels, compactness=compactness, sigma=sigma) segments = torch.from_numpy(segments).unsqueeze(0).unsqueeze(0).float() segments = F.interpolate(segments, size=(imgA.height, imgA.width), mode='nearest').long() # 应用超像素块范围到图像B imgB_np = np.array(imgB) for i in range(num_superpixels): mask = (segments == i) imgB_np[mask.expand(3, -1, -1)] = np.mean(imgB_np[mask.expand(3, -1, -1)], axis=1, keepdims=True) # 显示超像素分割图像 imgA_segments = np.zeros_like(imgA_np) for i in range(num_superpixels): mask = (segments == i) imgA_segments[mask.expand(3, -1, -1)] = np.random.randint(0, 255, size=(3,)) imgA_segments = Image.fromarray(imgA_segments.astype(np.uint8)) imgB_segments = Image.fromarray(imgB_np) # 显示图像 transforms.ToPILImage()(imgA).show() transforms.ToPILImage()(imgB).show() imgA_segments.show() imgB_segments.show()上述代码出现错误:RuntimeError: expand(CPUBoolType{[1, 1, 512, 512]}, size=[3, -1, -1]): the number of sizes provided (3) must be greater or equal to the number of dimensions in the tensor (4)

import numpy as np from scipy.stats import norm # Parameters S0 = 1.5 # initial FX rate U = 1.7 # upper barrier level L = 1.2 # lower barrier level X = 1.4 # strike price T = 1.0 # time to maturity r = 0.03 # risk-free rate rf = 0.0 # foreign interest rate sigma = 0.12 # volatility # Simulation settings M = 100000 # number of Monte Carlo simulations N = 252 # number of time steps # Time and step size dt = T / N t = np.linspace(0, T, N+1) # Simulate FX rates Z = np.random.standard_normal((M, N)) S = np.zeros((M, N+1)) S[:, 0] = S0 for i in range(N): S[:, i+1] = S[:, i] * np.exp((r-rf - 0.5*sigma**2)*dt + sigma*np.sqrt(dt)*Z[:, i]) # Compute option payoff payoff = np.zeros(M) for i in range(M): # Check if the option has knocked out if np.any((S[i, 21:126] > U) | (S[i, 201:231] < L) | (S[i, -1] < 1.3) | (S[i, -1] > 1.8)): payoff[i] = 0 else: payoff[i] = np.maximum(S[i, -1] - X, 0) # Compute option price and standard deviation using Monte Carlo simulation discount_factor = np.exp(-r*T) option_price = discount_factor * np.mean(payoff) std_dev = np.std(payoff) print("Option price:", option_price) print("Standard deviation:", std_dev) # Compute option delta using finite difference method delta = np.zeros(N+1) delta[0] = norm.cdf((np.log(S0/X) + (r-rf + 0.5*sigma**2)*T) / (sigma*np.sqrt(T))) for i in range(1, N+1): Si = S[:, i] Si_minus_1 = S[:, i-1] Ci = np.maximum(Si-X, 0) Ci_minus_1 = np.maximum(Si_minus_1-X, 0) delta[i] = np.mean((Ci - Ci_minus_1) / (Si - Si_minus_1)) * np.exp(-r*dt) print("Option delta:", delta[-1]) File "<ipython-input-2-57deb9637f96>", line 34, in <module> if np.any((S[i, 21:126] > U) | (S[i, 201:231] < L) | (S[i, -1] < 1.3) | (S[i, -1] > 1.8)): ValueError: operands could not be broadcast together with shapes (105,) (30,)

import torch import torchvision.transforms as transforms import numpy as np from skimage.segmentation import slic from skimage.segmentation import mark_boundaries from skimage.filters import sobel from skimage.color import rgb2gray from PIL import Image # 超像素数量 num_segments = 100 # 加载图像 image = Image.open('test.jpg') # 转换为 PyTorch 张量 transform = transforms.ToTensor() img_tensor = transform(image).unsqueeze(0) # 转换为 Numpy 数组 img_np = img_tensor.numpy().transpose(0, 2, 3, 1)[0] # 转换为灰度图像 gray_img = rgb2gray(img_np) # 使用 SLIC 超像素分割算法 segments = slic(img_np, n_segments=num_segments, compactness=10, sigma=1) # 绘制超像素边界线 edge_img = mark_boundaries(img_np, segments) # 转换为灰度图像 gray_edge_img = rgb2gray(edge_img) # 使用 Canny 边缘检测算法 edges = sobel(gray_edge_img) edge_map = edges > np.mean(edges) # 绘制超像素范围的线 line_map = np.zeros_like(gray_img) for i in range(num_segments): line_map[segments == i] = edge_map[segments == i].max() # 将线绘制到图像上 line_img = np.zeros_like(img_np) line_img[:, :, 0] = line_map line_img[:, :, 1] = line_map line_img[:, :, 2] = line_map result_img = img_np * (1 - line_img) + line_img * np.array([1, 0, 0]) # 显示结果 result_img = (result_img * 255).astype(np.uint8) result_img = Image.fromarray(result_img) result_img.show()上述代码出现问题:alueError: zero-size array to reduction operation maximum which has no identity

import numpy as npimport cv2# 读取图像img = cv2.imread('lena.png', 0)# 添加高斯噪声mean = 0var = 0.1sigma = var ** 0.5noise = np.random.normal(mean, sigma, img.shape)noisy_img = img + noise# 定义维纳滤波器函数def wiener_filter(img, psf, K=0.01): # 计算傅里叶变换 img_fft = np.fft.fft2(img) psf_fft = np.fft.fft2(psf) # 计算功率谱 img_power = np.abs(img_fft) ** 2 psf_power = np.abs(psf_fft) ** 2 # 计算信噪比 snr = img_power / (psf_power + K) # 计算滤波器 result_fft = img_fft * snr / psf_fft result = np.fft.ifft2(result_fft) # 返回滤波结果 return np.abs(result)# 定义维纳滤波器的卷积核kernel_size = 3kernel = np.ones((kernel_size, kernel_size)) / kernel_size ** 2# 计算图像的自相关函数acf = cv2.calcHist([img], [0], None, [256], [0, 256])# 计算维纳滤波器的卷积核gamma = 0.1alpha = 0.5beta = 1 - alpha - gammapsf = np.zeros((kernel_size, kernel_size))for i in range(kernel_size): for j in range(kernel_size): i_shift = i - kernel_size // 2 j_shift = j - kernel_size // 2 psf[i, j] = np.exp(-np.pi * ((i_shift ** 2 + j_shift ** 2) / (2 * alpha ** 2))) * np.cos(2 * np.pi * (i_shift + j_shift) / (2 * beta))psf = psf / np.sum(psf)# 对带噪声图像进行维纳滤波filtered_img = wiener_filter(noisy_img, psf)# 显示结果cv2.imshow('Original Image', img)cv2.imshow('Noisy Image', noisy_img)cv2.imshow('Filtered Image', filtered_img)cv2.waitKey(0)cv2.destroyAllWindows()这段代码报错为Traceback (most recent call last): File "<input>", line 1, in <module> File "D:\Pycharm\PyCharm 2020.3.5\plugins\python\helpers\pydev\_pydev_bundle\pydev_umd.py", line 197, in runfile pydev_imports.execfile(filename, global_vars, local_vars) # execute the script File "D:\Pycharm\PyCharm 2020.3.5\plugins\python\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile exec(compile(contents+"\n", file, 'exec'), glob, loc) File "E:/Python_project/class_/weinalvboqi.py", line 54, in <module> filtered_img = wiener_filter(noisy_img, psf) File "E:/Python_project/class_/weinalvboqi.py", line 25, in wiener_filter snr = img_power / (psf_power + K) ValueError: operands could not be broadcast together with shapes (1024,2800) (3,3)什么意思,如何修改

Here are the detail information provided in PPTs:The option is an exotic partial barrier option written on an FX rate. The current value of underlying FX rate S0 = 1.5 (i.e. 1.5 units of domestic buys 1 unit of foreign). It matures in one year, i.e. T = 1. The option knocks out, if the FX rate:1 is greater than an upper level U in the period between between 1 month’s time and 6 month’s time; or,2 is less than a lower level L in the period between 8th month and 11th month; or,3 lies outside the interval [1.3, 1.8] in the final month up to the end of year.If it has not been knocked out at the end of year, the owner has the option to buy 1 unit of foreign for X units of domestic, say X = 1.4, then, the payoff is max{0, ST − X }.We assume that, FX rate follows a geometric Brownian motion dSt = μSt dt + σSt dWt , (20) where under risk-neutrality μ = r − rf = 0.03 and σ = 0.12.To simulate path, we divide the time period [0, T ] into N small intervals of length ∆t = T /N, and discretize the SDE above by Euler approximation St +∆t − St = μSt ∆t + σSt √∆tZt , Zt ∼ N (0, 1). (21) The algorithm for pricing this barrier option by Monte Carlo simulation is as described as follows:1 Initialize S0;2 Take Si∆t as known, calculate S(i+1)∆t using equation the discretized SDE as above;3 If Si+1 hits any barrier, then set payoff to be 0 and stop iteration, otherwise, set payoff at time T to max{0, ST − X };4 Repeat the above steps for M times and get M payoffs;5 Calculate the average of M payoffs and discount at rate μ;6 Calculate the standard deviation of M payoffs.

最新推荐

recommend-type

pre_o_1csdn63m9a1bs0e1rr51niuu33e.a

pre_o_1csdn63m9a1bs0e1rr51niuu33e.a
recommend-type

matlab建立计算力学课程的笔记和文件.zip

matlab建立计算力学课程的笔记和文件.zip
recommend-type

FT-Prog-v3.12.38.643-FTD USB 工作模式设定及eprom读写

FT_Prog_v3.12.38.643--FTD USB 工作模式设定及eprom读写
recommend-type

zigbee-cluster-library-specification

最新的zigbee-cluster-library-specification说明文档。
recommend-type

管理建模和仿真的文件

管理Boualem Benatallah引用此版本:布阿利姆·贝纳塔拉。管理建模和仿真。约瑟夫-傅立叶大学-格勒诺布尔第一大学,1996年。法语。NNT:电话:00345357HAL ID:电话:00345357https://theses.hal.science/tel-003453572008年12月9日提交HAL是一个多学科的开放存取档案馆,用于存放和传播科学研究论文,无论它们是否被公开。论文可以来自法国或国外的教学和研究机构,也可以来自公共或私人研究中心。L’archive ouverte pluridisciplinaire
recommend-type

实现实时数据湖架构:Kafka与Hive集成

![实现实时数据湖架构:Kafka与Hive集成](https://img-blog.csdnimg.cn/img_convert/10eb2e6972b3b6086286fc64c0b3ee41.jpeg) # 1. 实时数据湖架构概述** 实时数据湖是一种现代数据管理架构,它允许企业以低延迟的方式收集、存储和处理大量数据。与传统数据仓库不同,实时数据湖不依赖于预先定义的模式,而是采用灵活的架构,可以处理各种数据类型和格式。这种架构为企业提供了以下优势: - **实时洞察:**实时数据湖允许企业访问最新的数据,从而做出更明智的决策。 - **数据民主化:**实时数据湖使各种利益相关者都可
recommend-type

2. 通过python绘制y=e-xsin(2πx)图像

可以使用matplotlib库来绘制这个函数的图像。以下是一段示例代码: ```python import numpy as np import matplotlib.pyplot as plt def func(x): return np.exp(-x) * np.sin(2 * np.pi * x) x = np.linspace(0, 5, 500) y = func(x) plt.plot(x, y) plt.xlabel('x') plt.ylabel('y') plt.title('y = e^{-x} sin(2πx)') plt.show() ``` 运行这段
recommend-type

JSBSim Reference Manual

JSBSim参考手册,其中包含JSBSim简介,JSBSim配置文件xml的编写语法,编程手册以及一些应用实例等。其中有部分内容还没有写完,估计有生之年很难看到完整版了,但是内容还是很有参考价值的。
recommend-type

"互动学习:行动中的多样性与论文攻读经历"

多样性她- 事实上SCI NCES你的时间表ECOLEDO C Tora SC和NCESPOUR l’Ingén学习互动,互动学习以行动为中心的强化学习学会互动,互动学习,以行动为中心的强化学习计算机科学博士论文于2021年9月28日在Villeneuve d'Asq公开支持马修·瑟林评审团主席法布里斯·勒菲弗尔阿维尼翁大学教授论文指导奥利维尔·皮耶昆谷歌研究教授:智囊团论文联合主任菲利普·普雷教授,大学。里尔/CRISTAL/因里亚报告员奥利维耶·西格德索邦大学报告员卢多维奇·德诺耶教授,Facebook /索邦大学审查员越南圣迈IMT Atlantic高级讲师邀请弗洛里安·斯特鲁布博士,Deepmind对于那些及时看到自己错误的人...3谢谢你首先,我要感谢我的两位博士生导师Olivier和Philippe。奥利维尔,"站在巨人的肩膀上"这句话对你来说完全有意义了。从科学上讲,你知道在这篇论文的(许多)错误中,你是我可以依
recommend-type

实现实时监控告警系统:Kafka与Grafana整合

![实现实时监控告警系统:Kafka与Grafana整合](https://imgconvert.csdnimg.cn/aHR0cHM6Ly9tbWJpei5xcGljLmNuL21tYml6X2pwZy9BVldpY3ladXVDbEZpY1pLWmw2bUVaWXFUcEdLT1VDdkxRSmQxZXB5R1lxaWNlUjA2c0hFek5Qc3FyRktudFF1VDMxQVl3QTRXV2lhSWFRMEFRc0I1cW1ZOGcvNjQw?x-oss-process=image/format,png) # 1.1 Kafka集群架构 Kafka集群由多个称为代理的服务器组成,这