Java实现线性表:增删改查操作

需积分: 6 0 下载量 177 浏览量 更新于2024-08-05 收藏 17KB TXT 举报
"这是一个Java实现的线性表,基于数组数据结构,包含了增、删、改、查等基本操作。" 在计算机科学中,线性表是一种基础的数据结构,它是由n(n>=0)个相同类型元素构成的有限序列。在这个Java实现中,线性表是通过数组来实现的,这使得我们可以直接访问数组中的任何元素,且具有较低的查找和修改时间复杂度。以下是代码中包含的主要知识点: 1. **数组长度计算**:`len()` 方法用于获取数组的长度。在Java中,数组的长度是固定的,可以通过 `arr.length` 获取。 2. **数组遍历**:`printList()` 方法遍历整个数组并打印所有元素。使用for循环遍历数组,当数组不为空时,从索引0开始直到最后一个元素。 3. **索引元素**:`printElement()` 方法返回数组中指定位置的元素。它首先检查索引是否在有效范围内,然后返回对应位置的元素。 4. **索引元素群体**:`printElements()` 方法用于获取数组中一段连续元素的子数组。它接受起始和结束索引,创建一个新的数组,并将原数组中对应位置的元素复制到新数组中。 5. **元素存在性判断**:`boolElement()` 方法检查数组中是否存在特定的元素。它遍历数组并与目标元素比较,如果找到匹配则返回true,否则返回false。 6. **元素群体存在性判断**:`boolElements()` 方法检查数组中是否存在一个元素群体。它接收一个元素列表,然后逐一检查这些元素是否都在原数组中。 7. **插入操作**:虽然没有直接提供,但在实际应用中,要在数组实现的线性表中插入元素,通常需要创建新的数组,将旧数组中的元素复制过来,再插入新元素,最后调整数组长度。 8. **删除操作**:同样,删除元素通常涉及到创建新的数组并跳过待删除元素,然后更新数组长度。 9. **修改操作**:修改数组中的元素可以直接通过数组索引来完成,例如 `arr[index] = newValue;`。 这个实现虽然简单,但涵盖了线性表的基本操作。在实际项目中,可能需要扩展这些功能,比如支持动态扩容的数组,或者使用链表实现线性表以提高插入和删除的效率。此外,为了使代码更健壮,还可以添加异常处理来应对无效输入或越界情况。

import numpy as np import matplotlib.pyplot as plt from sklearn import svm from sklearn.datasets import make_blobs from sklearn import model_selection from sklearn.metrics import f1_score def show_svm(a, b, bt): plt.figure(bt) plt.title('SVM with ' + bt) # 建立图像坐标 axis = plt.gca() plt.scatter(a[:, 0], a[:, 1], c=b, s=30) xlim = [a[:, 0].min(), a[:, 0].max()] ylim = [a[:, 1].min(), a[:, 1].max()] # 生成两个等差数列 xx = np.linspace(xlim[0], xlim[1], 50) yy = np.linspace(ylim[0], ylim[1], 50) X, Y = np.meshgrid(xx, yy) xy = np.vstack([X.ravel(), Y.ravel()]).T Z = clf.decision_function(xy).reshape(X.shape) # 画出分界线 axis.contour(X, Y, Z, colors='k', levels=[-1, 0, 1], alpha=0.5, linestyles=['--', '-', '--']) axis.scatter(clf.support_vectors_[:, 0], clf.support_vectors_[:, 1], s=200, linewidths=1, facecolors='none') if __name__ == '__main__': # data = np.loadtxt('separable_data.txt', delimiter=',') # data = np.loadtxt('non_separable_data.txt', delimiter=',') # data = np.loadtxt('banknote.txt', delimiter=',') data = np.loadtxt('ionosphere.txt', delimiter=',') # data = np.loadtxt('wdbc.txt', delimiter=',') X = data[:, 0:-1] y = data[:, -1] """标签中有一类标签为1""" y = y + 1 ymin = min(y) if not (1 in set(y)): ll = max(list(set(y))) + 1 for i in range(len(y)): if y[i] == ymin: y[i] = 1 # 建立一个线性核(多项式核)的SVM clf = svm.SVC(kernel='linear') clf.fit(X, y) """显示所有数据用于训练后的可视化结果""" show_svm(X, y, 'all dataset') """divide the data into two sections: training and test datasets""" X_train, X_test, y_train, y_test = model_selection.train_test_split(X, y, test_size=0.1, random_state=42) """training""" clf = svm.SVC(kernel='linear')#线性内核 # clf = svm.SVC(kernel='poly')# 多项式内核 # clf = svm.SVC(kernel='sigmoid')# Sigmoid内核 clf.fit(X_train, y_train) # show_svm(X_train, y_train, 'training dataset') """predict""" pred = clf.predict(X_test) pred = np.array(pred) y_test = np.array(y_test) print(f'SVM 的预测结果 f1-score:{f1_score(y_test, pred)}') # plt.show()结果与分析

2023-06-08 上传

import torch import torch.nn as nn import torch.optim as optim import numpy as np 定义基本循环神经网络模型 class RNNModel(nn.Module): def init(self, rnn_type, input_size, hidden_size, output_size, num_layers=1): super(RNNModel, self).init() self.rnn_type = rnn_type self.input_size = input_size self.hidden_size = hidden_size self.output_size = output_size self.num_layers = num_layers self.encoder = nn.Embedding(input_size, hidden_size) if rnn_type == 'RNN': self.rnn = nn.RNN(hidden_size, hidden_size, num_layers) elif rnn_type == 'GRU': self.rnn = nn.GRU(hidden_size, hidden_size, num_layers) self.decoder = nn.Linear(hidden_size, output_size) def forward(self, input, hidden): input = self.encoder(input) output, hidden = self.rnn(input, hidden) output = output.view(-1, self.hidden_size) output = self.decoder(output) return output, hidden def init_hidden(self, batch_size): if self.rnn_type == 'RNN': return torch.zeros(self.num_layers, batch_size, self.hidden_size) elif self.rnn_type == 'GRU': return torch.zeros(self.num_layers, batch_size, self.hidden_size) 定义数据集 with open('汉语音节表.txt', encoding='utf-8') as f: chars = f.readline() chars = list(chars) idx_to_char = list(set(chars)) char_to_idx = dict([(char, i) for i, char in enumerate(idx_to_char)]) corpus_indices = [char_to_idx[char] for char in chars] 定义超参数 input_size = len(idx_to_char) hidden_size = 256 output_size = len(idx_to_char) num_layers = 1 batch_size = 32 num_steps = 5 learning_rate = 0.01 num_epochs = 100 定义模型、损失函数和优化器 model = RNNModel('RNN', input_size, hidden_size, output_size, num_layers) criterion = nn.CrossEntropyLoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) 训练模型 for epoch in range(num_epochs): model.train() hidden = model.init_hidden(batch_size) loss = 0 for X, Y in data_iter_consecutive(corpus_indices, batch_size, num_steps): optimizer.zero_grad() hidden = hidden.detach() output, hidden = model(X, hidden) loss = criterion(output, Y.view(-1)) loss.backward() torch.nn.utils.clip_grad_norm_(model.parameters(), max_norm=1.0) optimizer.step() if epoch % 10 == 0: print(f"Epoch {epoch}, Loss: {loss.item()}")请正确缩进代码

2023-06-09 上传

import pandas as pd data = pd.read_csv(C:\Users\Administrator\Desktop\pythonsjwj\weibo_senti_100k.csv') data = data.dropna(); data.shape data.head() import jieba data['data_cut'] = data['review'].apply(lambda x: list(jieba.cut(x))) data.head() with open('stopword.txt','r',encoding = 'utf-8') as f: stop = f.readlines() import re stop = [re.sub(' |\n|\ufeff','',r) for r in stop] data['data_after'] = [[i for i in s if i not in stop] for s in data['data_cut']] data.head() w = [] for i in data['data_after']: w.extend(i) num_data = pd.DataFrame(pd.Series(w).value_counts()) num_data['id'] = list(range(1,len(num_data)+1)) a = lambda x:list(num_data['id'][x]) data['vec'] = data['data_after'].apply(a) data.head() from wordcloud import WordCloud import matplotlib.pyplot as plt num_words = [''.join(i) for i in data['data_after']] num_words = ''.join(num_words) num_words= re.sub(' ','',num_words) num = pd.Series(jieba.lcut(num_words)).value_counts() wc_pic = WordCloud(background_color='white',font_path=r'C:\Windows\Fonts\simhei.ttf').fit_words(num) plt.figure(figsize=(10,10)) plt.imshow(wc_pic) plt.axis('off') plt.show() from sklearn.model_selection import train_test_split from keras.preprocessing import sequence maxlen = 128 vec_data = list(sequence.pad_sequences(data['vec'],maxlen=maxlen)) x,xt,y,yt = train_test_split(vec_data,data['label'],test_size = 0.2,random_state = 123) import numpy as np x = np.array(list(x)) y = np.array(list(y)) xt = np.array(list(xt)) yt = np.array(list(yt)) x=x[:2000,:] y=y[:2000] xt=xt[:500,:] yt=yt[:500] from sklearn.svm import SVC clf = SVC(C=1, kernel = 'linear') clf.fit(x,y) from sklearn.metrics import classification_report test_pre = clf.predict(xt) report = classification_report(yt,test_pre) print(report) from keras.optimizers import SGD, RMSprop, Adagrad from keras.utils import np_utils from keras.models import Sequential from keras.layers.core import Dense, Dropout, Activation from keras.layers.embeddings import Embedding from keras.layers.recurrent import LSTM, GRU model = Sequential() model.add(Embedding(len(num_data['id'])+1,256)) model.add(Dense(32, activation='sigmoid', input_dim=100)) model.add(LSTM(128)) model.add(Dense(1)) model.add(Activation('sigmoid')) model.summary() import matplotlib.pyplot as plt import matplotlib.image as mpimg from keras.utils import plot_model plot_model(model,to_file='Lstm2.png',show_shapes=True) ls = mpimg.imread('Lstm2.png') plt.imshow(ls) plt.axis('off') plt.show() model.compile(loss='binary_crossentropy',optimizer='Adam',metrics=["accuracy"]) model.fit(x,y,validation_data=(x,y),epochs=15)

2023-06-10 上传