[英]Convert C++ method to C#
我有一個解密base64的方法,我需要將其轉換為c#,我嘗試使用C#的Convert.FromBase64String以及我在這里看到的其他一些技巧,但最終的字符串與結果不符C ++。
有人可以幫忙理解C ++代碼的這種平靜嗎?
輸入示例: lWXtYpNpwBBXoLmcuktDqg==
預期輸出: AC0105E92ED496ACD373D7496B76C1C8
使用Convert.FromBase64的輸出: 95-65-ED-62-93-69-C0-10-57-A0-B9-9C-BA-4B-43-AA
static BYTE BackFrom64[]={
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0x3e,0xff,0xff,0xff,0x3f,
0x34,0x35,0x36,0x37,0x38,0x39,0x3a,0x3b,
0x3c,0x3d,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0x00,0x01,0x02,0x03,0x04,0x05,0x06,
0x07,0x08,0x09,0x0a,0x0b,0x0c,0x0d,0x0e,
0x0f,0x10,0x11,0x12,0x13,0x14,0x15,0x16,
0x17,0x18,0x19,0xff,0xff,0xff,0xff,0xff,
0xff,0x1a,0x1b,0x1c,0x1d,0x1e,0x1f,0x20,
0x21,0x22,0x23,0x24,0x25,0x26,0x27,0x28,
0x29,0x2a,0x2b,0x2c,0x2d,0x2e,0x2f,0x30,
0x31,0x32,0x33,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff,
0xff,0xff,0xff,0xff,0xff,0xff,0xff,0xff
};
static BYTE * TabBase64=(BYTE *)"ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/";
int FromBase64(BYTE *Output,BYTE *Input,int BufSize,int MaxSize) {
int xut,txut;
int i,t,k,shift;
BYTE c;
dWORD dr,dr1;
xut=txut=t=0;
dr=0L;
for(i=0; i < BufSize; ++i) {
c=*Input++;
if(c == '=') {
break;
}
if((c=BackFrom64[c]) == 0xff) {
continue;
}
dr1=(dWORD)c;
shift=(32-(6*(t+1)));
dr1<<=shift;
dr|=dr1;
t++;
if(t == 4) {
t=0;
for(k=0; k < 3; ++k) {
dr1=dr;
dr1>>=(32-(8*(k+1)));
dr1&=0xff;
*Output++=(BYTE) dr1;
if(++txut >= MaxSize) {
return(-1);
}
}
dr=0;
}
}
if(t > 0) {
for(k=0; k < (t-1); ++k) {
dr1=dr;
dr1>>=(32-(8*(k+1)));
dr1&=0xff;
*Output++=(BYTE) dr1;
if(++txut >= MaxSize) {
return(-1);
}
}
}
return(txut);
}
只需很少的工作即可完成非常幼稚的轉換。
唯一的其他變化是計算未供給BackFrom64
陣列作為逆TabBase64
陣列,以及添加一個單元測試用於所提供的輸入和輸出。 如果它執行的不是標准算法,則我將整理計數器,其中一些是多余的(可能將i
和iInput
替換為foreach(var c in input)
using System;
using Microsoft.VisualStudio.TestTools.UnitTesting;
namespace SO
{
[TestClass]
public class UnitTest1
{
private const string TabBase64 = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/";
private static readonly byte[] BackFrom64 = CreateBackFrom64();
private static byte[] CreateBackFrom64()
{
var backFrom64 = new byte[128];
for (var i = 0; i < backFrom64.Length; ++i)
backFrom64[i] = 0xff;
for (byte i = 0; i < TabBase64.Length; i++)
{
var c = TabBase64[i];
backFrom64[c] = i;
}
return backFrom64;
}
int FromBase64(byte[] Output, string Input, int BufSize, int MaxSize)
{
int xut, txut;
int i, t, k, shift;
byte c;
int dr, dr1;
xut = txut = t = 0;
dr = 0;
var iInput = 0;
var iOutput = 0;
for (i = 0; i < BufSize; ++i)
{
c = (byte)Input[iInput++];
if (c == '=')
{
break;
}
if ((c = BackFrom64[c]) == 0xff)
{
continue;
}
dr1 = c;
shift = (32 - (6 * (t + 1)));
dr1 <<= shift;
dr |= dr1;
t++;
if (t == 4)
{
t = 0;
for (k = 0; k < 3; ++k)
{
dr1 = dr;
dr1 >>= (32 - (8 * (k + 1)));
dr1 &= 0xff;
Output[iOutput++] = (byte)dr1;
if (++txut >= MaxSize)
{
return (-1);
}
}
dr = 0;
}
}
if (t > 0)
{
for (k = 0; k < (t - 1); ++k)
{
dr1 = dr;
dr1 >>= (32 - (8 * (k + 1)));
dr1 &= 0xff;
Output[iOutput++] = (byte)dr1;
if (++txut >= MaxSize)
{
return (-1);
}
}
}
return (txut);
}
[TestMethod]
public void TestMethod1()
{
var input = @"lWXtYpNpwBBXoLmcuktDqg==";
var expectedOutput = "AC0105E92ED496ACD373D7496B76C1C8";
var converted = new byte[(input.Length * 2) / 3];
var result = FromBase64(converted, input, input.Length, int.MaxValue);
Assert.AreNotEqual(-1, result, "result is valid");
Assert.AreEqual(
BitConverter.ToString(converted).Replace("-",""),
BitConverter.ToString(Convert.FromBase64String(input)).Replace("-",""),
"Supplied algorithm matches standard algorithm" );
Assert.AreEqual(
expectedOutput,
BitConverter.ToString(converted).Replace("-",""),
"Supplied algorithm matches expected output" );
}
}
}
現在,第一個斷言通過了,因為移植的代碼返回的結果與Convert.FromBase64String
完全相同,但是第二個斷言失敗了,因為這不是您說的從C ++獲得的輸出。
由此我可以得出結論,造成這種差異的最可能原因是在調用C ++版本時, BackFrom64
的內容尚未正確初始化,並且包含垃圾值-編輯到OP中提供的BackFrom64
的值導致相同的輸出作為計算得出的結果,因此預期的結果正在發生其他變化。 編譯C ++版本並使用給定的輸入運行:
int main()
{
const char* input = "lWXtYpNpwBBXoLmcuktDqg==";
BYTE Input[100];
BYTE Output[100];
int length = strlen(input);
memcpy(Input, input, length);
int result = FromBase64(Output, Input, 100, length);
for (int i = 0; i < result; ++i)
std::cout << std::hex << (int)Output[i] << " ";
std::cout << "\n";
}
產生輸出
95 65 ed 62 93 69 c0 10 57 a0 b9 9c ba 4b 43 aa
這再次表明它與標准算法相同。
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.