1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900490149024903490449054906490749084909491049114912491349144915491649174918491949204921492249234924492549264927492849294930493149324933493449354936493749384939494049414942494349444945494649474948494949504951495249534954495549564957495849594960496149624963496449654966496749684969497049714972497349744975497649774978497949804981498249834984498549864987498849894990499149924993499449954996499749984999500050015002500350045005500650075008500950105011501250135014501550165017501850195020502150225023502450255026502750285029503050315032503350345035503650375038503950405041504250435044504550465047504850495050505150525053505450555056505750585059506050615062506350645065506650675068506950705071507250735074507550765077507850795080508150825083508450855086508750885089509050915092509350945095509650975098509951005101510251035104510551065107510851095110511151125113511451155116511751185119512051215122512351245125512651275128512951305131513251335134513551365137513851395140514151425143514451455146514751485149515051515152515351545155515651575158515951605161516251635164516551665167516851695170517151725173517451755176517751785179518051815182518351845185518651875188518951905191519251935194519551965197519851995200520152025203520452055206520752085209521052115212521352145215521652175218521952205221522252235224522552265227522852295230523152325233523452355236523752385239524052415242524352445245524652475248524952505251525252535254525552565257525852595260526152625263526452655266526752685269527052715272527352745275527652775278527952805281528252835284528552865287528852895290529152925293529452955296529752985299530053015302530353045305530653075308530953105311531253135314531553165317531853195320532153225323532453255326532753285329533053315332533353345335533653375338533953405341534253435344534553465347534853495350535153525353535453555356535753585359536053615362536353645365536653675368536953705371537253735374537553765377537853795380538153825383538453855386538753885389539053915392539353945395539653975398539954005401540254035404540554065407540854095410541154125413541454155416541754185419542054215422542354245425542654275428542954305431543254335434543554365437543854395440544154425443544454455446544754485449545054515452545354545455545654575458545954605461546254635464546554665467546854695470547154725473547454755476547754785479548054815482548354845485548654875488548954905491549254935494549554965497549854995500550155025503550455055506550755085509551055115512551355145515551655175518551955205521552255235524552555265527552855295530553155325533553455355536553755385539554055415542554355445545554655475548554955505551555255535554555555565557555855595560556155625563556455655566556755685569557055715572557355745575557655775578557955805581558255835584558555865587558855895590559155925593559455955596559755985599560056015602 |
- // SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
- /* Copyright (c) 2015 - 2021 Intel Corporation */
- #include <linux/etherdevice.h>
- #include "osdep.h"
- #include "hmc.h"
- #include "defs.h"
- #include "type.h"
- #include "ws.h"
- #include "protos.h"
- /**
- * irdma_get_qp_from_list - get next qp from a list
- * @head: Listhead of qp's
- * @qp: current qp
- */
- struct irdma_sc_qp *irdma_get_qp_from_list(struct list_head *head,
- struct irdma_sc_qp *qp)
- {
- struct list_head *lastentry;
- struct list_head *entry = NULL;
- if (list_empty(head))
- return NULL;
- if (!qp) {
- entry = head->next;
- } else {
- lastentry = &qp->list;
- entry = lastentry->next;
- if (entry == head)
- return NULL;
- }
- return container_of(entry, struct irdma_sc_qp, list);
- }
- /**
- * irdma_sc_suspend_resume_qps - suspend/resume all qp's on VSI
- * @vsi: the VSI struct pointer
- * @op: Set to IRDMA_OP_RESUME or IRDMA_OP_SUSPEND
- */
- void irdma_sc_suspend_resume_qps(struct irdma_sc_vsi *vsi, u8 op)
- {
- struct irdma_sc_qp *qp = NULL;
- u8 i;
- for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
- mutex_lock(&vsi->qos[i].qos_mutex);
- qp = irdma_get_qp_from_list(&vsi->qos[i].qplist, qp);
- while (qp) {
- if (op == IRDMA_OP_RESUME) {
- if (!qp->dev->ws_add(vsi, i)) {
- qp->qs_handle =
- vsi->qos[qp->user_pri].qs_handle;
- irdma_cqp_qp_suspend_resume(qp, op);
- } else {
- irdma_cqp_qp_suspend_resume(qp, op);
- irdma_modify_qp_to_err(qp);
- }
- } else if (op == IRDMA_OP_SUSPEND) {
- /* issue cqp suspend command */
- if (!irdma_cqp_qp_suspend_resume(qp, op))
- atomic_inc(&vsi->qp_suspend_reqs);
- }
- qp = irdma_get_qp_from_list(&vsi->qos[i].qplist, qp);
- }
- mutex_unlock(&vsi->qos[i].qos_mutex);
- }
- }
- static void irdma_set_qos_info(struct irdma_sc_vsi *vsi,
- struct irdma_l2params *l2p)
- {
- u8 i;
- vsi->qos_rel_bw = l2p->vsi_rel_bw;
- vsi->qos_prio_type = l2p->vsi_prio_type;
- vsi->dscp_mode = l2p->dscp_mode;
- if (l2p->dscp_mode) {
- memcpy(vsi->dscp_map, l2p->dscp_map, sizeof(vsi->dscp_map));
- for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++)
- l2p->up2tc[i] = i;
- }
- for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
- if (vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1)
- vsi->qos[i].qs_handle = l2p->qs_handle_list[i];
- vsi->qos[i].traffic_class = l2p->up2tc[i];
- vsi->qos[i].rel_bw =
- l2p->tc_info[vsi->qos[i].traffic_class].rel_bw;
- vsi->qos[i].prio_type =
- l2p->tc_info[vsi->qos[i].traffic_class].prio_type;
- vsi->qos[i].valid = false;
- }
- }
- /**
- * irdma_change_l2params - given the new l2 parameters, change all qp
- * @vsi: RDMA VSI pointer
- * @l2params: New parameters from l2
- */
- void irdma_change_l2params(struct irdma_sc_vsi *vsi,
- struct irdma_l2params *l2params)
- {
- if (l2params->mtu_changed) {
- vsi->mtu = l2params->mtu;
- if (vsi->ieq)
- irdma_reinitialize_ieq(vsi);
- }
- if (!l2params->tc_changed)
- return;
- vsi->tc_change_pending = false;
- irdma_set_qos_info(vsi, l2params);
- irdma_sc_suspend_resume_qps(vsi, IRDMA_OP_RESUME);
- }
- /**
- * irdma_qp_rem_qos - remove qp from qos lists during destroy qp
- * @qp: qp to be removed from qos
- */
- void irdma_qp_rem_qos(struct irdma_sc_qp *qp)
- {
- struct irdma_sc_vsi *vsi = qp->vsi;
- ibdev_dbg(to_ibdev(qp->dev),
- "DCB: DCB: Remove qp[%d] UP[%d] qset[%d] on_qoslist[%d]\n",
- qp->qp_uk.qp_id, qp->user_pri, qp->qs_handle,
- qp->on_qoslist);
- mutex_lock(&vsi->qos[qp->user_pri].qos_mutex);
- if (qp->on_qoslist) {
- qp->on_qoslist = false;
- list_del(&qp->list);
- }
- mutex_unlock(&vsi->qos[qp->user_pri].qos_mutex);
- }
- /**
- * irdma_qp_add_qos - called during setctx for qp to be added to qos
- * @qp: qp to be added to qos
- */
- void irdma_qp_add_qos(struct irdma_sc_qp *qp)
- {
- struct irdma_sc_vsi *vsi = qp->vsi;
- ibdev_dbg(to_ibdev(qp->dev),
- "DCB: DCB: Add qp[%d] UP[%d] qset[%d] on_qoslist[%d]\n",
- qp->qp_uk.qp_id, qp->user_pri, qp->qs_handle,
- qp->on_qoslist);
- mutex_lock(&vsi->qos[qp->user_pri].qos_mutex);
- if (!qp->on_qoslist) {
- list_add(&qp->list, &vsi->qos[qp->user_pri].qplist);
- qp->on_qoslist = true;
- qp->qs_handle = vsi->qos[qp->user_pri].qs_handle;
- }
- mutex_unlock(&vsi->qos[qp->user_pri].qos_mutex);
- }
- /**
- * irdma_sc_pd_init - initialize sc pd struct
- * @dev: sc device struct
- * @pd: sc pd ptr
- * @pd_id: pd_id for allocated pd
- * @abi_ver: User/Kernel ABI version
- */
- void irdma_sc_pd_init(struct irdma_sc_dev *dev, struct irdma_sc_pd *pd, u32 pd_id,
- int abi_ver)
- {
- pd->pd_id = pd_id;
- pd->abi_ver = abi_ver;
- pd->dev = dev;
- }
- /**
- * irdma_sc_add_arp_cache_entry - cqp wqe add arp cache entry
- * @cqp: struct for cqp hw
- * @info: arp entry information
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_add_arp_cache_entry(struct irdma_sc_cqp *cqp,
- struct irdma_add_arp_cache_entry_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 8, info->reach_max);
- set_64bit_val(wqe, 16, ether_addr_to_u64(info->mac_addr));
- hdr = info->arp_index |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MANAGE_ARP) |
- FIELD_PREP(IRDMA_CQPSQ_MAT_PERMANENT, (info->permanent ? 1 : 0)) |
- FIELD_PREP(IRDMA_CQPSQ_MAT_ENTRYVALID, 1) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: ARP_CACHE_ENTRY WQE", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_del_arp_cache_entry - dele arp cache entry
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @arp_index: arp index to delete arp entry
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_del_arp_cache_entry(struct irdma_sc_cqp *cqp, u64 scratch,
- u16 arp_index, bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- hdr = arp_index |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MANAGE_ARP) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: ARP_CACHE_DEL_ENTRY WQE",
- DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_manage_apbvt_entry - for adding and deleting apbvt entries
- * @cqp: struct for cqp hw
- * @info: info for apbvt entry to add or delete
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_manage_apbvt_entry(struct irdma_sc_cqp *cqp,
- struct irdma_apbvt_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, info->port);
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MANAGE_APBVT) |
- FIELD_PREP(IRDMA_CQPSQ_MAPT_ADDPORT, info->add) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: MANAGE_APBVT WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_manage_qhash_table_entry - manage quad hash entries
- * @cqp: struct for cqp hw
- * @info: info for quad hash to manage
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- *
- * This is called before connection establishment is started.
- * For passive connections, when listener is created, it will
- * call with entry type of IRDMA_QHASH_TYPE_TCP_SYN with local
- * ip address and tcp port. When SYN is received (passive
- * connections) or sent (active connections), this routine is
- * called with entry type of IRDMA_QHASH_TYPE_TCP_ESTABLISHED
- * and quad is passed in info.
- *
- * When iwarp connection is done and its state moves to RTS, the
- * quad hash entry in the hardware will point to iwarp's qp
- * number and requires no calls from the driver.
- */
- static int
- irdma_sc_manage_qhash_table_entry(struct irdma_sc_cqp *cqp,
- struct irdma_qhash_table_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 qw1 = 0;
- u64 qw2 = 0;
- u64 temp;
- struct irdma_sc_vsi *vsi = info->vsi;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 0, ether_addr_to_u64(info->mac_addr));
- qw1 = FIELD_PREP(IRDMA_CQPSQ_QHASH_QPN, info->qp_num) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_DEST_PORT, info->dest_port);
- if (info->ipv4_valid) {
- set_64bit_val(wqe, 48,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR3, info->dest_ip[0]));
- } else {
- set_64bit_val(wqe, 56,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR0, info->dest_ip[0]) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR1, info->dest_ip[1]));
- set_64bit_val(wqe, 48,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR2, info->dest_ip[2]) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR3, info->dest_ip[3]));
- }
- qw2 = FIELD_PREP(IRDMA_CQPSQ_QHASH_QS_HANDLE,
- vsi->qos[info->user_pri].qs_handle);
- if (info->vlan_valid)
- qw2 |= FIELD_PREP(IRDMA_CQPSQ_QHASH_VLANID, info->vlan_id);
- set_64bit_val(wqe, 16, qw2);
- if (info->entry_type == IRDMA_QHASH_TYPE_TCP_ESTABLISHED) {
- qw1 |= FIELD_PREP(IRDMA_CQPSQ_QHASH_SRC_PORT, info->src_port);
- if (!info->ipv4_valid) {
- set_64bit_val(wqe, 40,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR0, info->src_ip[0]) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR1, info->src_ip[1]));
- set_64bit_val(wqe, 32,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR2, info->src_ip[2]) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR3, info->src_ip[3]));
- } else {
- set_64bit_val(wqe, 32,
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ADDR3, info->src_ip[0]));
- }
- }
- set_64bit_val(wqe, 8, qw1);
- temp = FIELD_PREP(IRDMA_CQPSQ_QHASH_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_OPCODE,
- IRDMA_CQP_OP_MANAGE_QUAD_HASH_TABLE_ENTRY) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_MANAGE, info->manage) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_IPV4VALID, info->ipv4_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_VLANVALID, info->vlan_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QHASH_ENTRYTYPE, info->entry_type);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("WQE: MANAGE_QHASH WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_qp_init - initialize qp
- * @qp: sc qp
- * @info: initialization qp info
- */
- int irdma_sc_qp_init(struct irdma_sc_qp *qp, struct irdma_qp_init_info *info)
- {
- int ret_code;
- u32 pble_obj_cnt;
- u16 wqe_size;
- if (info->qp_uk_init_info.max_sq_frag_cnt >
- info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags ||
- info->qp_uk_init_info.max_rq_frag_cnt >
- info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags)
- return -EINVAL;
- qp->dev = info->pd->dev;
- qp->vsi = info->vsi;
- qp->ieq_qp = info->vsi->exception_lan_q;
- qp->sq_pa = info->sq_pa;
- qp->rq_pa = info->rq_pa;
- qp->hw_host_ctx_pa = info->host_ctx_pa;
- qp->q2_pa = info->q2_pa;
- qp->shadow_area_pa = info->shadow_area_pa;
- qp->q2_buf = info->q2;
- qp->pd = info->pd;
- qp->hw_host_ctx = info->host_ctx;
- info->qp_uk_init_info.wqe_alloc_db = qp->pd->dev->wqe_alloc_db;
- ret_code = irdma_uk_qp_init(&qp->qp_uk, &info->qp_uk_init_info);
- if (ret_code)
- return ret_code;
- qp->virtual_map = info->virtual_map;
- pble_obj_cnt = info->pd->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if ((info->virtual_map && info->sq_pa >= pble_obj_cnt) ||
- (info->virtual_map && info->rq_pa >= pble_obj_cnt))
- return -EINVAL;
- qp->llp_stream_handle = (void *)(-1);
- qp->hw_sq_size = irdma_get_encoded_wqe_size(qp->qp_uk.sq_ring.size,
- IRDMA_QUEUE_TYPE_SQ_RQ);
- ibdev_dbg(to_ibdev(qp->dev),
- "WQE: hw_sq_size[%04d] sq_ring.size[%04d]\n",
- qp->hw_sq_size, qp->qp_uk.sq_ring.size);
- if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1 && qp->pd->abi_ver > 4)
- wqe_size = IRDMA_WQE_SIZE_128;
- else
- ret_code = irdma_fragcnt_to_wqesize_rq(qp->qp_uk.max_rq_frag_cnt,
- &wqe_size);
- if (ret_code)
- return ret_code;
- qp->hw_rq_size = irdma_get_encoded_wqe_size(qp->qp_uk.rq_size *
- (wqe_size / IRDMA_QP_WQE_MIN_SIZE), IRDMA_QUEUE_TYPE_SQ_RQ);
- ibdev_dbg(to_ibdev(qp->dev),
- "WQE: hw_rq_size[%04d] qp_uk.rq_size[%04d] wqe_size[%04d]\n",
- qp->hw_rq_size, qp->qp_uk.rq_size, wqe_size);
- qp->sq_tph_val = info->sq_tph_val;
- qp->rq_tph_val = info->rq_tph_val;
- qp->sq_tph_en = info->sq_tph_en;
- qp->rq_tph_en = info->rq_tph_en;
- qp->rcv_tph_en = info->rcv_tph_en;
- qp->xmit_tph_en = info->xmit_tph_en;
- qp->qp_uk.first_sq_wq = info->qp_uk_init_info.first_sq_wq;
- qp->qs_handle = qp->vsi->qos[qp->user_pri].qs_handle;
- return 0;
- }
- /**
- * irdma_sc_qp_create - create qp
- * @qp: sc qp
- * @info: qp create info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_qp_create(struct irdma_sc_qp *qp, struct irdma_create_qp_info *info,
- u64 scratch, bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- cqp = qp->dev->cqp;
- if (qp->qp_uk.qp_id < cqp->dev->hw_attrs.min_hw_qp_id ||
- qp->qp_uk.qp_id >= cqp->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_QP].max_cnt)
- return -EINVAL;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, qp->hw_host_ctx_pa);
- set_64bit_val(wqe, 40, qp->shadow_area_pa);
- hdr = qp->qp_uk.qp_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_CREATE_QP) |
- FIELD_PREP(IRDMA_CQPSQ_QP_ORDVALID, (info->ord_valid ? 1 : 0)) |
- FIELD_PREP(IRDMA_CQPSQ_QP_TOECTXVALID, info->tcp_ctx_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_MACVALID, info->mac_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_QPTYPE, qp->qp_uk.qp_type) |
- FIELD_PREP(IRDMA_CQPSQ_QP_VQ, qp->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_QP_FORCELOOPBACK, info->force_lpb) |
- FIELD_PREP(IRDMA_CQPSQ_QP_CQNUMVALID, info->cq_num_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_ARPTABIDXVALID,
- info->arp_cache_idx_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_NEXTIWSTATE, info->next_iwarp_state) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QP_CREATE WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_qp_modify - modify qp cqp wqe
- * @qp: sc qp
- * @info: modify qp info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_qp_modify(struct irdma_sc_qp *qp, struct irdma_modify_qp_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- u8 term_actions = 0;
- u8 term_len = 0;
- cqp = qp->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- if (info->next_iwarp_state == IRDMA_QP_STATE_TERMINATE) {
- if (info->dont_send_fin)
- term_actions += IRDMAQP_TERM_SEND_TERM_ONLY;
- if (info->dont_send_term)
- term_actions += IRDMAQP_TERM_SEND_FIN_ONLY;
- if (term_actions == IRDMAQP_TERM_SEND_TERM_AND_FIN ||
- term_actions == IRDMAQP_TERM_SEND_TERM_ONLY)
- term_len = info->termlen;
- }
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMA_CQPSQ_QP_NEWMSS, info->new_mss) |
- FIELD_PREP(IRDMA_CQPSQ_QP_TERMLEN, term_len));
- set_64bit_val(wqe, 16, qp->hw_host_ctx_pa);
- set_64bit_val(wqe, 40, qp->shadow_area_pa);
- hdr = qp->qp_uk.qp_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MODIFY_QP) |
- FIELD_PREP(IRDMA_CQPSQ_QP_ORDVALID, info->ord_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_TOECTXVALID, info->tcp_ctx_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_CACHEDVARVALID,
- info->cached_var_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_VQ, qp->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_QP_FORCELOOPBACK, info->force_lpb) |
- FIELD_PREP(IRDMA_CQPSQ_QP_CQNUMVALID, info->cq_num_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_MACVALID, info->mac_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_QPTYPE, qp->qp_uk.qp_type) |
- FIELD_PREP(IRDMA_CQPSQ_QP_MSSCHANGE, info->mss_change) |
- FIELD_PREP(IRDMA_CQPSQ_QP_REMOVEHASHENTRY,
- info->remove_hash_idx) |
- FIELD_PREP(IRDMA_CQPSQ_QP_TERMACT, term_actions) |
- FIELD_PREP(IRDMA_CQPSQ_QP_RESETCON, info->reset_tcp_conn) |
- FIELD_PREP(IRDMA_CQPSQ_QP_ARPTABIDXVALID,
- info->arp_cache_idx_valid) |
- FIELD_PREP(IRDMA_CQPSQ_QP_NEXTIWSTATE, info->next_iwarp_state) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QP_MODIFY WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_qp_destroy - cqp destroy qp
- * @qp: sc qp
- * @scratch: u64 saved to be used during cqp completion
- * @remove_hash_idx: flag if to remove hash idx
- * @ignore_mw_bnd: memory window bind flag
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_qp_destroy(struct irdma_sc_qp *qp, u64 scratch,
- bool remove_hash_idx, bool ignore_mw_bnd, bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- cqp = qp->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, qp->hw_host_ctx_pa);
- set_64bit_val(wqe, 40, qp->shadow_area_pa);
- hdr = qp->qp_uk.qp_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DESTROY_QP) |
- FIELD_PREP(IRDMA_CQPSQ_QP_QPTYPE, qp->qp_uk.qp_type) |
- FIELD_PREP(IRDMA_CQPSQ_QP_IGNOREMWBOUND, ignore_mw_bnd) |
- FIELD_PREP(IRDMA_CQPSQ_QP_REMOVEHASHENTRY, remove_hash_idx) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QP_DESTROY WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_get_encoded_ird_size -
- * @ird_size: IRD size
- * The ird from the connection is rounded to a supported HW setting and then encoded
- * for ird_size field of qp_ctx. Consumers are expected to provide valid ird size based
- * on hardware attributes. IRD size defaults to a value of 4 in case of invalid input
- */
- static u8 irdma_sc_get_encoded_ird_size(u16 ird_size)
- {
- switch (ird_size ?
- roundup_pow_of_two(2 * ird_size) : 4) {
- case 256:
- return IRDMA_IRD_HW_SIZE_256;
- case 128:
- return IRDMA_IRD_HW_SIZE_128;
- case 64:
- case 32:
- return IRDMA_IRD_HW_SIZE_64;
- case 16:
- case 8:
- return IRDMA_IRD_HW_SIZE_16;
- case 4:
- default:
- break;
- }
- return IRDMA_IRD_HW_SIZE_4;
- }
- /**
- * irdma_sc_qp_setctx_roce - set qp's context
- * @qp: sc qp
- * @qp_ctx: context ptr
- * @info: ctx info
- */
- void irdma_sc_qp_setctx_roce(struct irdma_sc_qp *qp, __le64 *qp_ctx,
- struct irdma_qp_host_ctx_info *info)
- {
- struct irdma_roce_offload_info *roce_info;
- struct irdma_udp_offload_info *udp;
- u8 push_mode_en;
- u32 push_idx;
- roce_info = info->roce_info;
- udp = info->udp_info;
- qp->user_pri = info->user_pri;
- if (qp->push_idx == IRDMA_INVALID_PUSH_PAGE_INDEX) {
- push_mode_en = 0;
- push_idx = 0;
- } else {
- push_mode_en = 1;
- push_idx = qp->push_idx;
- }
- set_64bit_val(qp_ctx, 0,
- FIELD_PREP(IRDMAQPC_RQWQESIZE, qp->qp_uk.rq_wqe_size) |
- FIELD_PREP(IRDMAQPC_RCVTPHEN, qp->rcv_tph_en) |
- FIELD_PREP(IRDMAQPC_XMITTPHEN, qp->xmit_tph_en) |
- FIELD_PREP(IRDMAQPC_RQTPHEN, qp->rq_tph_en) |
- FIELD_PREP(IRDMAQPC_SQTPHEN, qp->sq_tph_en) |
- FIELD_PREP(IRDMAQPC_PPIDX, push_idx) |
- FIELD_PREP(IRDMAQPC_PMENA, push_mode_en) |
- FIELD_PREP(IRDMAQPC_PDIDXHI, roce_info->pd_id >> 16) |
- FIELD_PREP(IRDMAQPC_DC_TCP_EN, roce_info->dctcp_en) |
- FIELD_PREP(IRDMAQPC_ERR_RQ_IDX_VALID, roce_info->err_rq_idx_valid) |
- FIELD_PREP(IRDMAQPC_ISQP1, roce_info->is_qp1) |
- FIELD_PREP(IRDMAQPC_ROCE_TVER, roce_info->roce_tver) |
- FIELD_PREP(IRDMAQPC_IPV4, udp->ipv4) |
- FIELD_PREP(IRDMAQPC_INSERTVLANTAG, udp->insert_vlan_tag));
- set_64bit_val(qp_ctx, 8, qp->sq_pa);
- set_64bit_val(qp_ctx, 16, qp->rq_pa);
- if ((roce_info->dcqcn_en || roce_info->dctcp_en) &&
- !(udp->tos & 0x03))
- udp->tos |= ECN_CODE_PT_VAL;
- set_64bit_val(qp_ctx, 24,
- FIELD_PREP(IRDMAQPC_RQSIZE, qp->hw_rq_size) |
- FIELD_PREP(IRDMAQPC_SQSIZE, qp->hw_sq_size) |
- FIELD_PREP(IRDMAQPC_TTL, udp->ttl) | FIELD_PREP(IRDMAQPC_TOS, udp->tos) |
- FIELD_PREP(IRDMAQPC_SRCPORTNUM, udp->src_port) |
- FIELD_PREP(IRDMAQPC_DESTPORTNUM, udp->dst_port));
- set_64bit_val(qp_ctx, 32,
- FIELD_PREP(IRDMAQPC_DESTIPADDR2, udp->dest_ip_addr[2]) |
- FIELD_PREP(IRDMAQPC_DESTIPADDR3, udp->dest_ip_addr[3]));
- set_64bit_val(qp_ctx, 40,
- FIELD_PREP(IRDMAQPC_DESTIPADDR0, udp->dest_ip_addr[0]) |
- FIELD_PREP(IRDMAQPC_DESTIPADDR1, udp->dest_ip_addr[1]));
- set_64bit_val(qp_ctx, 48,
- FIELD_PREP(IRDMAQPC_SNDMSS, udp->snd_mss) |
- FIELD_PREP(IRDMAQPC_VLANTAG, udp->vlan_tag) |
- FIELD_PREP(IRDMAQPC_ARPIDX, udp->arp_idx));
- set_64bit_val(qp_ctx, 56,
- FIELD_PREP(IRDMAQPC_PKEY, roce_info->p_key) |
- FIELD_PREP(IRDMAQPC_PDIDX, roce_info->pd_id) |
- FIELD_PREP(IRDMAQPC_ACKCREDITS, roce_info->ack_credits) |
- FIELD_PREP(IRDMAQPC_FLOWLABEL, udp->flow_label));
- set_64bit_val(qp_ctx, 64,
- FIELD_PREP(IRDMAQPC_QKEY, roce_info->qkey) |
- FIELD_PREP(IRDMAQPC_DESTQP, roce_info->dest_qp));
- set_64bit_val(qp_ctx, 80,
- FIELD_PREP(IRDMAQPC_PSNNXT, udp->psn_nxt) |
- FIELD_PREP(IRDMAQPC_LSN, udp->lsn));
- set_64bit_val(qp_ctx, 88,
- FIELD_PREP(IRDMAQPC_EPSN, udp->epsn));
- set_64bit_val(qp_ctx, 96,
- FIELD_PREP(IRDMAQPC_PSNMAX, udp->psn_max) |
- FIELD_PREP(IRDMAQPC_PSNUNA, udp->psn_una));
- set_64bit_val(qp_ctx, 112,
- FIELD_PREP(IRDMAQPC_CWNDROCE, udp->cwnd));
- set_64bit_val(qp_ctx, 128,
- FIELD_PREP(IRDMAQPC_ERR_RQ_IDX, roce_info->err_rq_idx) |
- FIELD_PREP(IRDMAQPC_RNRNAK_THRESH, udp->rnr_nak_thresh) |
- FIELD_PREP(IRDMAQPC_REXMIT_THRESH, udp->rexmit_thresh) |
- FIELD_PREP(IRDMAQPC_RTOMIN, roce_info->rtomin));
- set_64bit_val(qp_ctx, 136,
- FIELD_PREP(IRDMAQPC_TXCQNUM, info->send_cq_num) |
- FIELD_PREP(IRDMAQPC_RXCQNUM, info->rcv_cq_num));
- set_64bit_val(qp_ctx, 144,
- FIELD_PREP(IRDMAQPC_STAT_INDEX, info->stats_idx));
- set_64bit_val(qp_ctx, 152, ether_addr_to_u64(roce_info->mac_addr) << 16);
- set_64bit_val(qp_ctx, 160,
- FIELD_PREP(IRDMAQPC_ORDSIZE, roce_info->ord_size) |
- FIELD_PREP(IRDMAQPC_IRDSIZE, irdma_sc_get_encoded_ird_size(roce_info->ird_size)) |
- FIELD_PREP(IRDMAQPC_WRRDRSPOK, roce_info->wr_rdresp_en) |
- FIELD_PREP(IRDMAQPC_RDOK, roce_info->rd_en) |
- FIELD_PREP(IRDMAQPC_USESTATSINSTANCE, info->stats_idx_valid) |
- FIELD_PREP(IRDMAQPC_BINDEN, roce_info->bind_en) |
- FIELD_PREP(IRDMAQPC_FASTREGEN, roce_info->fast_reg_en) |
- FIELD_PREP(IRDMAQPC_DCQCNENABLE, roce_info->dcqcn_en) |
- FIELD_PREP(IRDMAQPC_RCVNOICRC, roce_info->rcv_no_icrc) |
- FIELD_PREP(IRDMAQPC_FW_CC_ENABLE, roce_info->fw_cc_enable) |
- FIELD_PREP(IRDMAQPC_UDPRIVCQENABLE, roce_info->udprivcq_en) |
- FIELD_PREP(IRDMAQPC_PRIVEN, roce_info->priv_mode_en) |
- FIELD_PREP(IRDMAQPC_TIMELYENABLE, roce_info->timely_en));
- set_64bit_val(qp_ctx, 168,
- FIELD_PREP(IRDMAQPC_QPCOMPCTX, info->qp_compl_ctx));
- set_64bit_val(qp_ctx, 176,
- FIELD_PREP(IRDMAQPC_SQTPHVAL, qp->sq_tph_val) |
- FIELD_PREP(IRDMAQPC_RQTPHVAL, qp->rq_tph_val) |
- FIELD_PREP(IRDMAQPC_QSHANDLE, qp->qs_handle));
- set_64bit_val(qp_ctx, 184,
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR3, udp->local_ipaddr[3]) |
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR2, udp->local_ipaddr[2]));
- set_64bit_val(qp_ctx, 192,
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR1, udp->local_ipaddr[1]) |
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR0, udp->local_ipaddr[0]));
- set_64bit_val(qp_ctx, 200,
- FIELD_PREP(IRDMAQPC_THIGH, roce_info->t_high) |
- FIELD_PREP(IRDMAQPC_TLOW, roce_info->t_low));
- set_64bit_val(qp_ctx, 208,
- FIELD_PREP(IRDMAQPC_REMENDPOINTIDX, info->rem_endpoint_idx));
- print_hex_dump_debug("WQE: QP_HOST CTX WQE", DUMP_PREFIX_OFFSET, 16,
- 8, qp_ctx, IRDMA_QP_CTX_SIZE, false);
- }
- /* irdma_sc_alloc_local_mac_entry - allocate a mac entry
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_alloc_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch,
- bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_ALLOCATE_LOC_MAC_TABLE_ENTRY) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: ALLOCATE_LOCAL_MAC WQE",
- DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_add_local_mac_entry - add mac enry
- * @cqp: struct for cqp hw
- * @info:mac addr info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_add_local_mac_entry(struct irdma_sc_cqp *cqp,
- struct irdma_local_mac_entry_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 header;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 32, ether_addr_to_u64(info->mac_addr));
- header = FIELD_PREP(IRDMA_CQPSQ_MLM_TABLEIDX, info->entry_idx) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_MANAGE_LOC_MAC_TABLE) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, header);
- print_hex_dump_debug("WQE: ADD_LOCAL_MAC WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_del_local_mac_entry - cqp wqe to dele local mac
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @entry_idx: index of mac entry
- * @ignore_ref_count: to force mac adde delete
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_del_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch,
- u16 entry_idx, u8 ignore_ref_count,
- bool post_sq)
- {
- __le64 *wqe;
- u64 header;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- header = FIELD_PREP(IRDMA_CQPSQ_MLM_TABLEIDX, entry_idx) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_MANAGE_LOC_MAC_TABLE) |
- FIELD_PREP(IRDMA_CQPSQ_MLM_FREEENTRY, 1) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_MLM_IGNORE_REF_CNT, ignore_ref_count);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, header);
- print_hex_dump_debug("WQE: DEL_LOCAL_MAC_IPADDR WQE",
- DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_qp_setctx - set qp's context
- * @qp: sc qp
- * @qp_ctx: context ptr
- * @info: ctx info
- */
- void irdma_sc_qp_setctx(struct irdma_sc_qp *qp, __le64 *qp_ctx,
- struct irdma_qp_host_ctx_info *info)
- {
- struct irdma_iwarp_offload_info *iw;
- struct irdma_tcp_offload_info *tcp;
- struct irdma_sc_dev *dev;
- u8 push_mode_en;
- u32 push_idx;
- u64 qw0, qw3, qw7 = 0, qw16 = 0;
- u64 mac = 0;
- iw = info->iwarp_info;
- tcp = info->tcp_info;
- dev = qp->dev;
- if (iw->rcv_mark_en) {
- qp->pfpdu.marker_len = 4;
- qp->pfpdu.rcv_start_seq = tcp->rcv_nxt;
- }
- qp->user_pri = info->user_pri;
- if (qp->push_idx == IRDMA_INVALID_PUSH_PAGE_INDEX) {
- push_mode_en = 0;
- push_idx = 0;
- } else {
- push_mode_en = 1;
- push_idx = qp->push_idx;
- }
- qw0 = FIELD_PREP(IRDMAQPC_RQWQESIZE, qp->qp_uk.rq_wqe_size) |
- FIELD_PREP(IRDMAQPC_RCVTPHEN, qp->rcv_tph_en) |
- FIELD_PREP(IRDMAQPC_XMITTPHEN, qp->xmit_tph_en) |
- FIELD_PREP(IRDMAQPC_RQTPHEN, qp->rq_tph_en) |
- FIELD_PREP(IRDMAQPC_SQTPHEN, qp->sq_tph_en) |
- FIELD_PREP(IRDMAQPC_PPIDX, push_idx) |
- FIELD_PREP(IRDMAQPC_PMENA, push_mode_en);
- set_64bit_val(qp_ctx, 8, qp->sq_pa);
- set_64bit_val(qp_ctx, 16, qp->rq_pa);
- qw3 = FIELD_PREP(IRDMAQPC_RQSIZE, qp->hw_rq_size) |
- FIELD_PREP(IRDMAQPC_SQSIZE, qp->hw_sq_size);
- if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1)
- qw3 |= FIELD_PREP(IRDMAQPC_GEN1_SRCMACADDRIDX,
- qp->src_mac_addr_idx);
- set_64bit_val(qp_ctx, 136,
- FIELD_PREP(IRDMAQPC_TXCQNUM, info->send_cq_num) |
- FIELD_PREP(IRDMAQPC_RXCQNUM, info->rcv_cq_num));
- set_64bit_val(qp_ctx, 168,
- FIELD_PREP(IRDMAQPC_QPCOMPCTX, info->qp_compl_ctx));
- set_64bit_val(qp_ctx, 176,
- FIELD_PREP(IRDMAQPC_SQTPHVAL, qp->sq_tph_val) |
- FIELD_PREP(IRDMAQPC_RQTPHVAL, qp->rq_tph_val) |
- FIELD_PREP(IRDMAQPC_QSHANDLE, qp->qs_handle) |
- FIELD_PREP(IRDMAQPC_EXCEPTION_LAN_QUEUE, qp->ieq_qp));
- if (info->iwarp_info_valid) {
- qw0 |= FIELD_PREP(IRDMAQPC_DDP_VER, iw->ddp_ver) |
- FIELD_PREP(IRDMAQPC_RDMAP_VER, iw->rdmap_ver) |
- FIELD_PREP(IRDMAQPC_DC_TCP_EN, iw->dctcp_en) |
- FIELD_PREP(IRDMAQPC_ECN_EN, iw->ecn_en) |
- FIELD_PREP(IRDMAQPC_IBRDENABLE, iw->ib_rd_en) |
- FIELD_PREP(IRDMAQPC_PDIDXHI, iw->pd_id >> 16) |
- FIELD_PREP(IRDMAQPC_ERR_RQ_IDX_VALID,
- iw->err_rq_idx_valid);
- qw7 |= FIELD_PREP(IRDMAQPC_PDIDX, iw->pd_id);
- qw16 |= FIELD_PREP(IRDMAQPC_ERR_RQ_IDX, iw->err_rq_idx) |
- FIELD_PREP(IRDMAQPC_RTOMIN, iw->rtomin);
- set_64bit_val(qp_ctx, 144,
- FIELD_PREP(IRDMAQPC_Q2ADDR, qp->q2_pa >> 8) |
- FIELD_PREP(IRDMAQPC_STAT_INDEX, info->stats_idx));
- if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2)
- mac = ether_addr_to_u64(iw->mac_addr);
- set_64bit_val(qp_ctx, 152,
- mac << 16 | FIELD_PREP(IRDMAQPC_LASTBYTESENT, iw->last_byte_sent));
- set_64bit_val(qp_ctx, 160,
- FIELD_PREP(IRDMAQPC_ORDSIZE, iw->ord_size) |
- FIELD_PREP(IRDMAQPC_IRDSIZE, irdma_sc_get_encoded_ird_size(iw->ird_size)) |
- FIELD_PREP(IRDMAQPC_WRRDRSPOK, iw->wr_rdresp_en) |
- FIELD_PREP(IRDMAQPC_RDOK, iw->rd_en) |
- FIELD_PREP(IRDMAQPC_SNDMARKERS, iw->snd_mark_en) |
- FIELD_PREP(IRDMAQPC_BINDEN, iw->bind_en) |
- FIELD_PREP(IRDMAQPC_FASTREGEN, iw->fast_reg_en) |
- FIELD_PREP(IRDMAQPC_PRIVEN, iw->priv_mode_en) |
- FIELD_PREP(IRDMAQPC_USESTATSINSTANCE, info->stats_idx_valid) |
- FIELD_PREP(IRDMAQPC_IWARPMODE, 1) |
- FIELD_PREP(IRDMAQPC_RCVMARKERS, iw->rcv_mark_en) |
- FIELD_PREP(IRDMAQPC_ALIGNHDRS, iw->align_hdrs) |
- FIELD_PREP(IRDMAQPC_RCVNOMPACRC, iw->rcv_no_mpa_crc) |
- FIELD_PREP(IRDMAQPC_RCVMARKOFFSET, iw->rcv_mark_offset || !tcp ? iw->rcv_mark_offset : tcp->rcv_nxt) |
- FIELD_PREP(IRDMAQPC_SNDMARKOFFSET, iw->snd_mark_offset || !tcp ? iw->snd_mark_offset : tcp->snd_nxt) |
- FIELD_PREP(IRDMAQPC_TIMELYENABLE, iw->timely_en));
- }
- if (info->tcp_info_valid) {
- qw0 |= FIELD_PREP(IRDMAQPC_IPV4, tcp->ipv4) |
- FIELD_PREP(IRDMAQPC_NONAGLE, tcp->no_nagle) |
- FIELD_PREP(IRDMAQPC_INSERTVLANTAG,
- tcp->insert_vlan_tag) |
- FIELD_PREP(IRDMAQPC_TIMESTAMP, tcp->time_stamp) |
- FIELD_PREP(IRDMAQPC_LIMIT, tcp->cwnd_inc_limit) |
- FIELD_PREP(IRDMAQPC_DROPOOOSEG, tcp->drop_ooo_seg) |
- FIELD_PREP(IRDMAQPC_DUPACK_THRESH, tcp->dup_ack_thresh);
- if ((iw->ecn_en || iw->dctcp_en) && !(tcp->tos & 0x03))
- tcp->tos |= ECN_CODE_PT_VAL;
- qw3 |= FIELD_PREP(IRDMAQPC_TTL, tcp->ttl) |
- FIELD_PREP(IRDMAQPC_AVOIDSTRETCHACK, tcp->avoid_stretch_ack) |
- FIELD_PREP(IRDMAQPC_TOS, tcp->tos) |
- FIELD_PREP(IRDMAQPC_SRCPORTNUM, tcp->src_port) |
- FIELD_PREP(IRDMAQPC_DESTPORTNUM, tcp->dst_port);
- if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) {
- qw3 |= FIELD_PREP(IRDMAQPC_GEN1_SRCMACADDRIDX, tcp->src_mac_addr_idx);
- qp->src_mac_addr_idx = tcp->src_mac_addr_idx;
- }
- set_64bit_val(qp_ctx, 32,
- FIELD_PREP(IRDMAQPC_DESTIPADDR2, tcp->dest_ip_addr[2]) |
- FIELD_PREP(IRDMAQPC_DESTIPADDR3, tcp->dest_ip_addr[3]));
- set_64bit_val(qp_ctx, 40,
- FIELD_PREP(IRDMAQPC_DESTIPADDR0, tcp->dest_ip_addr[0]) |
- FIELD_PREP(IRDMAQPC_DESTIPADDR1, tcp->dest_ip_addr[1]));
- set_64bit_val(qp_ctx, 48,
- FIELD_PREP(IRDMAQPC_SNDMSS, tcp->snd_mss) |
- FIELD_PREP(IRDMAQPC_SYN_RST_HANDLING, tcp->syn_rst_handling) |
- FIELD_PREP(IRDMAQPC_VLANTAG, tcp->vlan_tag) |
- FIELD_PREP(IRDMAQPC_ARPIDX, tcp->arp_idx));
- qw7 |= FIELD_PREP(IRDMAQPC_FLOWLABEL, tcp->flow_label) |
- FIELD_PREP(IRDMAQPC_WSCALE, tcp->wscale) |
- FIELD_PREP(IRDMAQPC_IGNORE_TCP_OPT,
- tcp->ignore_tcp_opt) |
- FIELD_PREP(IRDMAQPC_IGNORE_TCP_UNS_OPT,
- tcp->ignore_tcp_uns_opt) |
- FIELD_PREP(IRDMAQPC_TCPSTATE, tcp->tcp_state) |
- FIELD_PREP(IRDMAQPC_RCVSCALE, tcp->rcv_wscale) |
- FIELD_PREP(IRDMAQPC_SNDSCALE, tcp->snd_wscale);
- set_64bit_val(qp_ctx, 72,
- FIELD_PREP(IRDMAQPC_TIMESTAMP_RECENT, tcp->time_stamp_recent) |
- FIELD_PREP(IRDMAQPC_TIMESTAMP_AGE, tcp->time_stamp_age));
- set_64bit_val(qp_ctx, 80,
- FIELD_PREP(IRDMAQPC_SNDNXT, tcp->snd_nxt) |
- FIELD_PREP(IRDMAQPC_SNDWND, tcp->snd_wnd));
- set_64bit_val(qp_ctx, 88,
- FIELD_PREP(IRDMAQPC_RCVNXT, tcp->rcv_nxt) |
- FIELD_PREP(IRDMAQPC_RCVWND, tcp->rcv_wnd));
- set_64bit_val(qp_ctx, 96,
- FIELD_PREP(IRDMAQPC_SNDMAX, tcp->snd_max) |
- FIELD_PREP(IRDMAQPC_SNDUNA, tcp->snd_una));
- set_64bit_val(qp_ctx, 104,
- FIELD_PREP(IRDMAQPC_SRTT, tcp->srtt) |
- FIELD_PREP(IRDMAQPC_RTTVAR, tcp->rtt_var));
- set_64bit_val(qp_ctx, 112,
- FIELD_PREP(IRDMAQPC_SSTHRESH, tcp->ss_thresh) |
- FIELD_PREP(IRDMAQPC_CWND, tcp->cwnd));
- set_64bit_val(qp_ctx, 120,
- FIELD_PREP(IRDMAQPC_SNDWL1, tcp->snd_wl1) |
- FIELD_PREP(IRDMAQPC_SNDWL2, tcp->snd_wl2));
- qw16 |= FIELD_PREP(IRDMAQPC_MAXSNDWND, tcp->max_snd_window) |
- FIELD_PREP(IRDMAQPC_REXMIT_THRESH, tcp->rexmit_thresh);
- set_64bit_val(qp_ctx, 184,
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR3, tcp->local_ipaddr[3]) |
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR2, tcp->local_ipaddr[2]));
- set_64bit_val(qp_ctx, 192,
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR1, tcp->local_ipaddr[1]) |
- FIELD_PREP(IRDMAQPC_LOCAL_IPADDR0, tcp->local_ipaddr[0]));
- set_64bit_val(qp_ctx, 200,
- FIELD_PREP(IRDMAQPC_THIGH, iw->t_high) |
- FIELD_PREP(IRDMAQPC_TLOW, iw->t_low));
- set_64bit_val(qp_ctx, 208,
- FIELD_PREP(IRDMAQPC_REMENDPOINTIDX, info->rem_endpoint_idx));
- }
- set_64bit_val(qp_ctx, 0, qw0);
- set_64bit_val(qp_ctx, 24, qw3);
- set_64bit_val(qp_ctx, 56, qw7);
- set_64bit_val(qp_ctx, 128, qw16);
- print_hex_dump_debug("WQE: QP_HOST CTX", DUMP_PREFIX_OFFSET, 16, 8,
- qp_ctx, IRDMA_QP_CTX_SIZE, false);
- }
- /**
- * irdma_sc_alloc_stag - mr stag alloc
- * @dev: sc device struct
- * @info: stag info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_alloc_stag(struct irdma_sc_dev *dev,
- struct irdma_allocate_stag_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- enum irdma_page_size page_size;
- if (!info->total_len && !info->all_memory)
- return -EINVAL;
- if (info->page_size == 0x40000000)
- page_size = IRDMA_PAGE_SIZE_1G;
- else if (info->page_size == 0x200000)
- page_size = IRDMA_PAGE_SIZE_2M;
- else
- page_size = IRDMA_PAGE_SIZE_4K;
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 8,
- FLD_LS_64(dev, info->pd_id, IRDMA_CQPSQ_STAG_PDID) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_STAGLEN, info->total_len));
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_STAG_IDX, info->stag_idx));
- set_64bit_val(wqe, 40,
- FIELD_PREP(IRDMA_CQPSQ_STAG_HMCFNIDX, info->hmc_fcn_index));
- if (info->chunk_size)
- set_64bit_val(wqe, 48,
- FIELD_PREP(IRDMA_CQPSQ_STAG_FIRSTPMPBLIDX, info->first_pm_pbl_idx));
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_ALLOC_STAG) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_MR, 1) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_ARIGHTS, info->access_rights) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_HPAGESIZE, page_size) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_REMACCENABLED, info->remote_access) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_USEHMCFNIDX, info->use_hmc_fcn_index) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_USEPFRID, info->use_pf_rid) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: ALLOC_STAG WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_mr_reg_non_shared - non-shared mr registration
- * @dev: sc device struct
- * @info: mr info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_mr_reg_non_shared(struct irdma_sc_dev *dev,
- struct irdma_reg_ns_stag_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 fbo;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- u32 pble_obj_cnt;
- bool remote_access;
- u8 addr_type;
- enum irdma_page_size page_size;
- if (!info->total_len && !info->all_memory)
- return -EINVAL;
- if (info->page_size == 0x40000000)
- page_size = IRDMA_PAGE_SIZE_1G;
- else if (info->page_size == 0x200000)
- page_size = IRDMA_PAGE_SIZE_2M;
- else if (info->page_size == 0x1000)
- page_size = IRDMA_PAGE_SIZE_4K;
- else
- return -EINVAL;
- if (info->access_rights & (IRDMA_ACCESS_FLAGS_REMOTEREAD_ONLY |
- IRDMA_ACCESS_FLAGS_REMOTEWRITE_ONLY))
- remote_access = true;
- else
- remote_access = false;
- pble_obj_cnt = dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->chunk_size && info->first_pm_pbl_index >= pble_obj_cnt)
- return -EINVAL;
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- fbo = info->va & (info->page_size - 1);
- set_64bit_val(wqe, 0,
- (info->addr_type == IRDMA_ADDR_TYPE_VA_BASED ?
- info->va : fbo));
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMA_CQPSQ_STAG_STAGLEN, info->total_len) |
- FLD_LS_64(dev, info->pd_id, IRDMA_CQPSQ_STAG_PDID));
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_STAG_KEY, info->stag_key) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_IDX, info->stag_idx));
- if (!info->chunk_size) {
- set_64bit_val(wqe, 32, info->reg_addr_pa);
- set_64bit_val(wqe, 48, 0);
- } else {
- set_64bit_val(wqe, 32, 0);
- set_64bit_val(wqe, 48,
- FIELD_PREP(IRDMA_CQPSQ_STAG_FIRSTPMPBLIDX, info->first_pm_pbl_index));
- }
- set_64bit_val(wqe, 40, info->hmc_fcn_index);
- set_64bit_val(wqe, 56, 0);
- addr_type = (info->addr_type == IRDMA_ADDR_TYPE_VA_BASED) ? 1 : 0;
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_REG_MR) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_MR, 1) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_HPAGESIZE, page_size) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_ARIGHTS, info->access_rights) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_REMACCENABLED, remote_access) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_VABASEDTO, addr_type) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_USEHMCFNIDX, info->use_hmc_fcn_index) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_USEPFRID, info->use_pf_rid) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: MR_REG_NS WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_dealloc_stag - deallocate stag
- * @dev: sc device struct
- * @info: dealloc stag info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_dealloc_stag(struct irdma_sc_dev *dev,
- struct irdma_dealloc_stag_info *info,
- u64 scratch, bool post_sq)
- {
- u64 hdr;
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 8,
- FLD_LS_64(dev, info->pd_id, IRDMA_CQPSQ_STAG_PDID));
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_STAG_IDX, info->stag_idx));
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DEALLOC_STAG) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_MR, info->mr) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: DEALLOC_STAG WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_mw_alloc - mw allocate
- * @dev: sc device struct
- * @info: memory window allocation information
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_mw_alloc(struct irdma_sc_dev *dev,
- struct irdma_mw_alloc_info *info, u64 scratch,
- bool post_sq)
- {
- u64 hdr;
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 8,
- FLD_LS_64(dev, info->pd_id, IRDMA_CQPSQ_STAG_PDID));
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_STAG_IDX, info->mw_stag_index));
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_ALLOC_STAG) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_MWTYPE, info->mw_wide) |
- FIELD_PREP(IRDMA_CQPSQ_STAG_MW1_BIND_DONT_VLDT_KEY,
- info->mw1_bind_dont_vldt_key) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: MW_ALLOC WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_mr_fast_register - Posts RDMA fast register mr WR to iwarp qp
- * @qp: sc qp struct
- * @info: fast mr info
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_mr_fast_register(struct irdma_sc_qp *qp,
- struct irdma_fast_reg_stag_info *info,
- bool post_sq)
- {
- u64 temp, hdr;
- __le64 *wqe;
- u32 wqe_idx;
- enum irdma_page_size page_size;
- struct irdma_post_sq_info sq_info = {};
- if (info->page_size == 0x40000000)
- page_size = IRDMA_PAGE_SIZE_1G;
- else if (info->page_size == 0x200000)
- page_size = IRDMA_PAGE_SIZE_2M;
- else
- page_size = IRDMA_PAGE_SIZE_4K;
- sq_info.wr_id = info->wr_id;
- sq_info.signaled = info->signaled;
- sq_info.push_wqe = info->push_wqe;
- wqe = irdma_qp_get_next_send_wqe(&qp->qp_uk, &wqe_idx,
- IRDMA_QP_WQE_MIN_QUANTA, 0, &sq_info);
- if (!wqe)
- return -ENOMEM;
- irdma_clr_wqes(&qp->qp_uk, wqe_idx);
- ibdev_dbg(to_ibdev(qp->dev),
- "MR: wr_id[%llxh] wqe_idx[%04d] location[%p]\n",
- info->wr_id, wqe_idx,
- &qp->qp_uk.sq_wrtrk_array[wqe_idx].wrid);
- temp = (info->addr_type == IRDMA_ADDR_TYPE_VA_BASED) ?
- (uintptr_t)info->va : info->fbo;
- set_64bit_val(wqe, 0, temp);
- temp = FIELD_GET(IRDMAQPSQ_FIRSTPMPBLIDXHI,
- info->first_pm_pbl_index >> 16);
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMAQPSQ_FIRSTPMPBLIDXHI, temp) |
- FIELD_PREP(IRDMAQPSQ_PBLADDR >> IRDMA_HW_PAGE_SHIFT, info->reg_addr_pa));
- set_64bit_val(wqe, 16,
- info->total_len |
- FIELD_PREP(IRDMAQPSQ_FIRSTPMPBLIDXLO, info->first_pm_pbl_index));
- hdr = FIELD_PREP(IRDMAQPSQ_STAGKEY, info->stag_key) |
- FIELD_PREP(IRDMAQPSQ_STAGINDEX, info->stag_idx) |
- FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_FAST_REGISTER) |
- FIELD_PREP(IRDMAQPSQ_LPBLSIZE, info->chunk_size) |
- FIELD_PREP(IRDMAQPSQ_HPAGESIZE, page_size) |
- FIELD_PREP(IRDMAQPSQ_STAGRIGHTS, info->access_rights) |
- FIELD_PREP(IRDMAQPSQ_VABASEDTO, info->addr_type) |
- FIELD_PREP(IRDMAQPSQ_PUSHWQE, (sq_info.push_wqe ? 1 : 0)) |
- FIELD_PREP(IRDMAQPSQ_READFENCE, info->read_fence) |
- FIELD_PREP(IRDMAQPSQ_LOCALFENCE, info->local_fence) |
- FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: FAST_REG WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_QP_WQE_MIN_SIZE, false);
- if (sq_info.push_wqe) {
- irdma_qp_push_wqe(&qp->qp_uk, wqe, IRDMA_QP_WQE_MIN_QUANTA,
- wqe_idx, post_sq);
- } else {
- if (post_sq)
- irdma_uk_qp_post_wr(&qp->qp_uk);
- }
- return 0;
- }
- /**
- * irdma_sc_gen_rts_ae - request AE generated after RTS
- * @qp: sc qp struct
- */
- static void irdma_sc_gen_rts_ae(struct irdma_sc_qp *qp)
- {
- __le64 *wqe;
- u64 hdr;
- struct irdma_qp_uk *qp_uk;
- qp_uk = &qp->qp_uk;
- wqe = qp_uk->sq_base[1].elem;
- hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_NOP) |
- FIELD_PREP(IRDMAQPSQ_LOCALFENCE, 1) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("QP: NOP W/LOCAL FENCE WQE", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_QP_WQE_MIN_SIZE, false);
- wqe = qp_uk->sq_base[2].elem;
- hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_GEN_RTS_AE) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("QP: CONN EST WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_QP_WQE_MIN_SIZE, false);
- }
- /**
- * irdma_sc_send_lsmm - send last streaming mode message
- * @qp: sc qp struct
- * @lsmm_buf: buffer with lsmm message
- * @size: size of lsmm buffer
- * @stag: stag of lsmm buffer
- */
- void irdma_sc_send_lsmm(struct irdma_sc_qp *qp, void *lsmm_buf, u32 size,
- irdma_stag stag)
- {
- __le64 *wqe;
- u64 hdr;
- struct irdma_qp_uk *qp_uk;
- qp_uk = &qp->qp_uk;
- wqe = qp_uk->sq_base->elem;
- set_64bit_val(wqe, 0, (uintptr_t)lsmm_buf);
- if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) {
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_LEN, size) |
- FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_STAG, stag));
- } else {
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMAQPSQ_FRAG_LEN, size) |
- FIELD_PREP(IRDMAQPSQ_FRAG_STAG, stag) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity));
- }
- set_64bit_val(wqe, 16, 0);
- hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_RDMA_SEND) |
- FIELD_PREP(IRDMAQPSQ_STREAMMODE, 1) |
- FIELD_PREP(IRDMAQPSQ_WAITFORRCVPDU, 1) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: SEND_LSMM WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_QP_WQE_MIN_SIZE, false);
- if (qp->dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE)
- irdma_sc_gen_rts_ae(qp);
- }
- /**
- * irdma_sc_send_rtt - send last read0 or write0
- * @qp: sc qp struct
- * @read: Do read0 or write0
- */
- void irdma_sc_send_rtt(struct irdma_sc_qp *qp, bool read)
- {
- __le64 *wqe;
- u64 hdr;
- struct irdma_qp_uk *qp_uk;
- qp_uk = &qp->qp_uk;
- wqe = qp_uk->sq_base->elem;
- set_64bit_val(wqe, 0, 0);
- set_64bit_val(wqe, 16, 0);
- if (read) {
- if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) {
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_STAG, 0xabcd));
- } else {
- set_64bit_val(wqe, 8,
- (u64)0xabcd | FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity));
- }
- hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, 0x1234) |
- FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_RDMA_READ) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- } else {
- if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) {
- set_64bit_val(wqe, 8, 0);
- } else {
- set_64bit_val(wqe, 8,
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity));
- }
- hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_RDMA_WRITE) |
- FIELD_PREP(IRDMAQPSQ_VALID, qp->qp_uk.swqe_polarity);
- }
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: RTR WQE", DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_QP_WQE_MIN_SIZE, false);
- if (qp->dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE)
- irdma_sc_gen_rts_ae(qp);
- }
- /**
- * irdma_iwarp_opcode - determine if incoming is rdma layer
- * @info: aeq info for the packet
- * @pkt: packet for error
- */
- static u32 irdma_iwarp_opcode(struct irdma_aeqe_info *info, u8 *pkt)
- {
- __be16 *mpa;
- u32 opcode = 0xffffffff;
- if (info->q2_data_written) {
- mpa = (__be16 *)pkt;
- opcode = ntohs(mpa[1]) & 0xf;
- }
- return opcode;
- }
- /**
- * irdma_locate_mpa - return pointer to mpa in the pkt
- * @pkt: packet with data
- */
- static u8 *irdma_locate_mpa(u8 *pkt)
- {
- /* skip over ethernet header */
- pkt += IRDMA_MAC_HLEN;
- /* Skip over IP and TCP headers */
- pkt += 4 * (pkt[0] & 0x0f);
- pkt += 4 * ((pkt[12] >> 4) & 0x0f);
- return pkt;
- }
- /**
- * irdma_bld_termhdr_ctrl - setup terminate hdr control fields
- * @qp: sc qp ptr for pkt
- * @hdr: term hdr
- * @opcode: flush opcode for termhdr
- * @layer_etype: error layer + error type
- * @err: error cod ein the header
- */
- static void irdma_bld_termhdr_ctrl(struct irdma_sc_qp *qp,
- struct irdma_terminate_hdr *hdr,
- enum irdma_flush_opcode opcode,
- u8 layer_etype, u8 err)
- {
- qp->flush_code = opcode;
- hdr->layer_etype = layer_etype;
- hdr->error_code = err;
- }
- /**
- * irdma_bld_termhdr_ddp_rdma - setup ddp and rdma hdrs in terminate hdr
- * @pkt: ptr to mpa in offending pkt
- * @hdr: term hdr
- * @copy_len: offending pkt length to be copied to term hdr
- * @is_tagged: DDP tagged or untagged
- */
- static void irdma_bld_termhdr_ddp_rdma(u8 *pkt, struct irdma_terminate_hdr *hdr,
- int *copy_len, u8 *is_tagged)
- {
- u16 ddp_seg_len;
- ddp_seg_len = ntohs(*(__be16 *)pkt);
- if (ddp_seg_len) {
- *copy_len = 2;
- hdr->hdrct = DDP_LEN_FLAG;
- if (pkt[2] & 0x80) {
- *is_tagged = 1;
- if (ddp_seg_len >= TERM_DDP_LEN_TAGGED) {
- *copy_len += TERM_DDP_LEN_TAGGED;
- hdr->hdrct |= DDP_HDR_FLAG;
- }
- } else {
- if (ddp_seg_len >= TERM_DDP_LEN_UNTAGGED) {
- *copy_len += TERM_DDP_LEN_UNTAGGED;
- hdr->hdrct |= DDP_HDR_FLAG;
- }
- if (ddp_seg_len >= (TERM_DDP_LEN_UNTAGGED + TERM_RDMA_LEN) &&
- ((pkt[3] & RDMA_OPCODE_M) == RDMA_READ_REQ_OPCODE)) {
- *copy_len += TERM_RDMA_LEN;
- hdr->hdrct |= RDMA_HDR_FLAG;
- }
- }
- }
- }
- /**
- * irdma_bld_terminate_hdr - build terminate message header
- * @qp: qp associated with received terminate AE
- * @info: the struct contiaing AE information
- */
- static int irdma_bld_terminate_hdr(struct irdma_sc_qp *qp,
- struct irdma_aeqe_info *info)
- {
- u8 *pkt = qp->q2_buf + Q2_BAD_FRAME_OFFSET;
- int copy_len = 0;
- u8 is_tagged = 0;
- u32 opcode;
- struct irdma_terminate_hdr *termhdr;
- termhdr = (struct irdma_terminate_hdr *)qp->q2_buf;
- memset(termhdr, 0, Q2_BAD_FRAME_OFFSET);
- if (info->q2_data_written) {
- pkt = irdma_locate_mpa(pkt);
- irdma_bld_termhdr_ddp_rdma(pkt, termhdr, ©_len, &is_tagged);
- }
- opcode = irdma_iwarp_opcode(info, pkt);
- qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC;
- qp->sq_flush_code = info->sq;
- qp->rq_flush_code = info->rq;
- switch (info->ae_id) {
- case IRDMA_AE_AMP_UNALLOCATED_STAG:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- if (opcode == IRDMA_OP_TYPE_RDMA_WRITE)
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_PROT_ERR,
- (LAYER_DDP << 4) | DDP_TAGGED_BUF,
- DDP_TAGGED_INV_STAG);
- else
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_INV_STAG);
- break;
- case IRDMA_AE_AMP_BOUNDS_VIOLATION:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- if (info->q2_data_written)
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_PROT_ERR,
- (LAYER_DDP << 4) | DDP_TAGGED_BUF,
- DDP_TAGGED_BOUNDS);
- else
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_INV_BOUNDS);
- break;
- case IRDMA_AE_AMP_BAD_PD:
- switch (opcode) {
- case IRDMA_OP_TYPE_RDMA_WRITE:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_PROT_ERR,
- (LAYER_DDP << 4) | DDP_TAGGED_BUF,
- DDP_TAGGED_UNASSOC_STAG);
- break;
- case IRDMA_OP_TYPE_SEND_INV:
- case IRDMA_OP_TYPE_SEND_SOL_INV:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_CANT_INV_STAG);
- break;
- default:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_UNASSOC_STAG);
- }
- break;
- case IRDMA_AE_AMP_INVALID_STAG:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_INV_STAG);
- break;
- case IRDMA_AE_AMP_BAD_QP:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_LOC_QP_OP_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_QN);
- break;
- case IRDMA_AE_AMP_BAD_STAG_KEY:
- case IRDMA_AE_AMP_BAD_STAG_INDEX:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- switch (opcode) {
- case IRDMA_OP_TYPE_SEND_INV:
- case IRDMA_OP_TYPE_SEND_SOL_INV:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_OP_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_OP,
- RDMAP_CANT_INV_STAG);
- break;
- default:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_OP,
- RDMAP_INV_STAG);
- }
- break;
- case IRDMA_AE_AMP_RIGHTS_VIOLATION:
- case IRDMA_AE_AMP_INVALIDATE_NO_REMOTE_ACCESS_RIGHTS:
- case IRDMA_AE_PRIV_OPERATION_DENIED:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_ACCESS);
- break;
- case IRDMA_AE_AMP_TO_WRAP:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_ACCESS_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_PROT,
- RDMAP_TO_WRAP);
- break;
- case IRDMA_AE_LLP_RECEIVED_MPA_CRC_ERROR:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_MPA << 4) | DDP_LLP, MPA_CRC);
- break;
- case IRDMA_AE_LLP_SEGMENT_TOO_SMALL:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_LOC_LEN_ERR,
- (LAYER_DDP << 4) | DDP_CATASTROPHIC,
- DDP_CATASTROPHIC_LOCAL);
- break;
- case IRDMA_AE_LCE_QP_CATASTROPHIC:
- case IRDMA_AE_DDP_NO_L_BIT:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_FATAL_ERR,
- (LAYER_DDP << 4) | DDP_CATASTROPHIC,
- DDP_CATASTROPHIC_LOCAL);
- break;
- case IRDMA_AE_DDP_INVALID_MSN_GAP_IN_MSN:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_MSN_RANGE);
- break;
- case IRDMA_AE_DDP_UBE_DDP_MESSAGE_TOO_LONG_FOR_AVAILABLE_BUFFER:
- qp->event_type = IRDMA_QP_EVENT_ACCESS_ERR;
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_LOC_LEN_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_TOO_LONG);
- break;
- case IRDMA_AE_DDP_UBE_INVALID_DDP_VERSION:
- if (is_tagged)
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_DDP << 4) | DDP_TAGGED_BUF,
- DDP_TAGGED_INV_DDP_VER);
- else
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_DDP_VER);
- break;
- case IRDMA_AE_DDP_UBE_INVALID_MO:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_MO);
- break;
- case IRDMA_AE_DDP_UBE_INVALID_MSN_NO_BUFFER_AVAILABLE:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_REM_OP_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_MSN_NO_BUF);
- break;
- case IRDMA_AE_DDP_UBE_INVALID_QN:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_DDP << 4) | DDP_UNTAGGED_BUF,
- DDP_UNTAGGED_INV_QN);
- break;
- case IRDMA_AE_RDMAP_ROE_INVALID_RDMAP_VERSION:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_GENERAL_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_OP,
- RDMAP_INV_RDMAP_VER);
- break;
- default:
- irdma_bld_termhdr_ctrl(qp, termhdr, FLUSH_FATAL_ERR,
- (LAYER_RDMA << 4) | RDMAP_REMOTE_OP,
- RDMAP_UNSPECIFIED);
- break;
- }
- if (copy_len)
- memcpy(termhdr + 1, pkt, copy_len);
- return sizeof(struct irdma_terminate_hdr) + copy_len;
- }
- /**
- * irdma_terminate_send_fin() - Send fin for terminate message
- * @qp: qp associated with received terminate AE
- */
- void irdma_terminate_send_fin(struct irdma_sc_qp *qp)
- {
- irdma_term_modify_qp(qp, IRDMA_QP_STATE_TERMINATE,
- IRDMAQP_TERM_SEND_FIN_ONLY, 0);
- }
- /**
- * irdma_terminate_connection() - Bad AE and send terminate to remote QP
- * @qp: qp associated with received terminate AE
- * @info: the struct contiaing AE information
- */
- void irdma_terminate_connection(struct irdma_sc_qp *qp,
- struct irdma_aeqe_info *info)
- {
- u8 termlen = 0;
- if (qp->term_flags & IRDMA_TERM_SENT)
- return;
- termlen = irdma_bld_terminate_hdr(qp, info);
- irdma_terminate_start_timer(qp);
- qp->term_flags |= IRDMA_TERM_SENT;
- irdma_term_modify_qp(qp, IRDMA_QP_STATE_TERMINATE,
- IRDMAQP_TERM_SEND_TERM_ONLY, termlen);
- }
- /**
- * irdma_terminate_received - handle terminate received AE
- * @qp: qp associated with received terminate AE
- * @info: the struct contiaing AE information
- */
- void irdma_terminate_received(struct irdma_sc_qp *qp,
- struct irdma_aeqe_info *info)
- {
- u8 *pkt = qp->q2_buf + Q2_BAD_FRAME_OFFSET;
- __be32 *mpa;
- u8 ddp_ctl;
- u8 rdma_ctl;
- u16 aeq_id = 0;
- struct irdma_terminate_hdr *termhdr;
- mpa = (__be32 *)irdma_locate_mpa(pkt);
- if (info->q2_data_written) {
- /* did not validate the frame - do it now */
- ddp_ctl = (ntohl(mpa[0]) >> 8) & 0xff;
- rdma_ctl = ntohl(mpa[0]) & 0xff;
- if ((ddp_ctl & 0xc0) != 0x40)
- aeq_id = IRDMA_AE_LCE_QP_CATASTROPHIC;
- else if ((ddp_ctl & 0x03) != 1)
- aeq_id = IRDMA_AE_DDP_UBE_INVALID_DDP_VERSION;
- else if (ntohl(mpa[2]) != 2)
- aeq_id = IRDMA_AE_DDP_UBE_INVALID_QN;
- else if (ntohl(mpa[3]) != 1)
- aeq_id = IRDMA_AE_DDP_INVALID_MSN_GAP_IN_MSN;
- else if (ntohl(mpa[4]) != 0)
- aeq_id = IRDMA_AE_DDP_UBE_INVALID_MO;
- else if ((rdma_ctl & 0xc0) != 0x40)
- aeq_id = IRDMA_AE_RDMAP_ROE_INVALID_RDMAP_VERSION;
- info->ae_id = aeq_id;
- if (info->ae_id) {
- /* Bad terminate recvd - send back a terminate */
- irdma_terminate_connection(qp, info);
- return;
- }
- }
- qp->term_flags |= IRDMA_TERM_RCVD;
- qp->event_type = IRDMA_QP_EVENT_CATASTROPHIC;
- termhdr = (struct irdma_terminate_hdr *)&mpa[5];
- if (termhdr->layer_etype == RDMAP_REMOTE_PROT ||
- termhdr->layer_etype == RDMAP_REMOTE_OP) {
- irdma_terminate_done(qp, 0);
- } else {
- irdma_terminate_start_timer(qp);
- irdma_terminate_send_fin(qp);
- }
- }
- static int irdma_null_ws_add(struct irdma_sc_vsi *vsi, u8 user_pri)
- {
- return 0;
- }
- static void irdma_null_ws_remove(struct irdma_sc_vsi *vsi, u8 user_pri)
- {
- /* do nothing */
- }
- static void irdma_null_ws_reset(struct irdma_sc_vsi *vsi)
- {
- /* do nothing */
- }
- /**
- * irdma_sc_vsi_init - Init the vsi structure
- * @vsi: pointer to vsi structure to initialize
- * @info: the info used to initialize the vsi struct
- */
- void irdma_sc_vsi_init(struct irdma_sc_vsi *vsi,
- struct irdma_vsi_init_info *info)
- {
- int i;
- vsi->dev = info->dev;
- vsi->back_vsi = info->back_vsi;
- vsi->register_qset = info->register_qset;
- vsi->unregister_qset = info->unregister_qset;
- vsi->mtu = info->params->mtu;
- vsi->exception_lan_q = info->exception_lan_q;
- vsi->vsi_idx = info->pf_data_vsi_num;
- if (vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1)
- vsi->fcn_id = info->dev->hmc_fn_id;
- irdma_set_qos_info(vsi, info->params);
- for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
- mutex_init(&vsi->qos[i].qos_mutex);
- INIT_LIST_HEAD(&vsi->qos[i].qplist);
- }
- if (vsi->register_qset) {
- vsi->dev->ws_add = irdma_ws_add;
- vsi->dev->ws_remove = irdma_ws_remove;
- vsi->dev->ws_reset = irdma_ws_reset;
- } else {
- vsi->dev->ws_add = irdma_null_ws_add;
- vsi->dev->ws_remove = irdma_null_ws_remove;
- vsi->dev->ws_reset = irdma_null_ws_reset;
- }
- }
- /**
- * irdma_get_fcn_id - Return the function id
- * @vsi: pointer to the vsi
- */
- static u8 irdma_get_fcn_id(struct irdma_sc_vsi *vsi)
- {
- struct irdma_stats_inst_info stats_info = {};
- struct irdma_sc_dev *dev = vsi->dev;
- u8 fcn_id = IRDMA_INVALID_FCN_ID;
- u8 start_idx, max_stats, i;
- if (dev->hw_attrs.uk_attrs.hw_rev != IRDMA_GEN_1) {
- if (!irdma_cqp_stats_inst_cmd(vsi, IRDMA_OP_STATS_ALLOCATE,
- &stats_info))
- return stats_info.stats_idx;
- }
- start_idx = 1;
- max_stats = 16;
- for (i = start_idx; i < max_stats; i++)
- if (!dev->fcn_id_array[i]) {
- fcn_id = i;
- dev->fcn_id_array[i] = true;
- break;
- }
- return fcn_id;
- }
- /**
- * irdma_vsi_stats_init - Initialize the vsi statistics
- * @vsi: pointer to the vsi structure
- * @info: The info structure used for initialization
- */
- int irdma_vsi_stats_init(struct irdma_sc_vsi *vsi,
- struct irdma_vsi_stats_info *info)
- {
- u8 fcn_id = info->fcn_id;
- struct irdma_dma_mem *stats_buff_mem;
- vsi->pestat = info->pestat;
- vsi->pestat->hw = vsi->dev->hw;
- vsi->pestat->vsi = vsi;
- stats_buff_mem = &vsi->pestat->gather_info.stats_buff_mem;
- stats_buff_mem->size = ALIGN(IRDMA_GATHER_STATS_BUF_SIZE * 2, 1);
- stats_buff_mem->va = dma_alloc_coherent(vsi->pestat->hw->device,
- stats_buff_mem->size,
- &stats_buff_mem->pa,
- GFP_KERNEL);
- if (!stats_buff_mem->va)
- return -ENOMEM;
- vsi->pestat->gather_info.gather_stats_va = stats_buff_mem->va;
- vsi->pestat->gather_info.last_gather_stats_va =
- (void *)((uintptr_t)stats_buff_mem->va +
- IRDMA_GATHER_STATS_BUF_SIZE);
- irdma_hw_stats_start_timer(vsi);
- if (info->alloc_fcn_id)
- fcn_id = irdma_get_fcn_id(vsi);
- if (fcn_id == IRDMA_INVALID_FCN_ID)
- goto stats_error;
- vsi->stats_fcn_id_alloc = info->alloc_fcn_id;
- vsi->fcn_id = fcn_id;
- if (info->alloc_fcn_id) {
- vsi->pestat->gather_info.use_stats_inst = true;
- vsi->pestat->gather_info.stats_inst_index = fcn_id;
- }
- return 0;
- stats_error:
- dma_free_coherent(vsi->pestat->hw->device, stats_buff_mem->size,
- stats_buff_mem->va, stats_buff_mem->pa);
- stats_buff_mem->va = NULL;
- return -EIO;
- }
- /**
- * irdma_vsi_stats_free - Free the vsi stats
- * @vsi: pointer to the vsi structure
- */
- void irdma_vsi_stats_free(struct irdma_sc_vsi *vsi)
- {
- struct irdma_stats_inst_info stats_info = {};
- u8 fcn_id = vsi->fcn_id;
- struct irdma_sc_dev *dev = vsi->dev;
- if (dev->hw_attrs.uk_attrs.hw_rev != IRDMA_GEN_1) {
- if (vsi->stats_fcn_id_alloc) {
- stats_info.stats_idx = vsi->fcn_id;
- irdma_cqp_stats_inst_cmd(vsi, IRDMA_OP_STATS_FREE,
- &stats_info);
- }
- } else {
- if (vsi->stats_fcn_id_alloc &&
- fcn_id < vsi->dev->hw_attrs.max_stat_inst)
- vsi->dev->fcn_id_array[fcn_id] = false;
- }
- if (!vsi->pestat)
- return;
- irdma_hw_stats_stop_timer(vsi);
- dma_free_coherent(vsi->pestat->hw->device,
- vsi->pestat->gather_info.stats_buff_mem.size,
- vsi->pestat->gather_info.stats_buff_mem.va,
- vsi->pestat->gather_info.stats_buff_mem.pa);
- vsi->pestat->gather_info.stats_buff_mem.va = NULL;
- }
- /**
- * irdma_get_encoded_wqe_size - given wq size, returns hardware encoded size
- * @wqsize: size of the wq (sq, rq) to encoded_size
- * @queue_type: queue type selected for the calculation algorithm
- */
- u8 irdma_get_encoded_wqe_size(u32 wqsize, enum irdma_queue_type queue_type)
- {
- u8 encoded_size = 0;
- /* cqp sq's hw coded value starts from 1 for size of 4
- * while it starts from 0 for qp' wq's.
- */
- if (queue_type == IRDMA_QUEUE_TYPE_CQP)
- encoded_size = 1;
- wqsize >>= 2;
- while (wqsize >>= 1)
- encoded_size++;
- return encoded_size;
- }
- /**
- * irdma_sc_gather_stats - collect the statistics
- * @cqp: struct for cqp hw
- * @info: gather stats info structure
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_gather_stats(struct irdma_sc_cqp *cqp,
- struct irdma_stats_gather_info *info,
- u64 scratch)
- {
- __le64 *wqe;
- u64 temp;
- if (info->stats_buff_mem.size < IRDMA_GATHER_STATS_BUF_SIZE)
- return -ENOMEM;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 40,
- FIELD_PREP(IRDMA_CQPSQ_STATS_HMC_FCN_INDEX, info->hmc_fcn_index));
- set_64bit_val(wqe, 32, info->stats_buff_mem.pa);
- temp = FIELD_PREP(IRDMA_CQPSQ_STATS_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_USE_INST, info->use_stats_inst) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_INST_INDEX,
- info->stats_inst_index) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_USE_HMC_FCN_INDEX,
- info->use_hmc_fcn_index) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_OP, IRDMA_CQP_OP_GATHER_STATS);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("STATS: GATHER_STATS WQE", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- ibdev_dbg(to_ibdev(cqp->dev),
- "STATS: CQP SQ head 0x%x tail 0x%x size 0x%x\n",
- cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size);
- return 0;
- }
- /**
- * irdma_sc_manage_stats_inst - allocate or free stats instance
- * @cqp: struct for cqp hw
- * @info: stats info structure
- * @alloc: alloc vs. delete flag
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_manage_stats_inst(struct irdma_sc_cqp *cqp,
- struct irdma_stats_inst_info *info,
- bool alloc, u64 scratch)
- {
- __le64 *wqe;
- u64 temp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 40,
- FIELD_PREP(IRDMA_CQPSQ_STATS_HMC_FCN_INDEX, info->hmc_fn_id));
- temp = FIELD_PREP(IRDMA_CQPSQ_STATS_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_ALLOC_INST, alloc) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_USE_HMC_FCN_INDEX,
- info->use_hmc_fcn_index) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_INST_INDEX, info->stats_idx) |
- FIELD_PREP(IRDMA_CQPSQ_STATS_OP, IRDMA_CQP_OP_MANAGE_STATS);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("WQE: MANAGE_STATS WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_set_up_map - set the up map table
- * @cqp: struct for cqp hw
- * @info: User priority map info
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_set_up_map(struct irdma_sc_cqp *cqp,
- struct irdma_up_info *info, u64 scratch)
- {
- __le64 *wqe;
- u64 temp = 0;
- int i;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++)
- temp |= (u64)info->map[i] << (i * 8);
- set_64bit_val(wqe, 0, temp);
- set_64bit_val(wqe, 40,
- FIELD_PREP(IRDMA_CQPSQ_UP_CNPOVERRIDE, info->cnp_up_override) |
- FIELD_PREP(IRDMA_CQPSQ_UP_HMCFCNIDX, info->hmc_fcn_idx));
- temp = FIELD_PREP(IRDMA_CQPSQ_UP_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_UP_USEVLAN, info->use_vlan) |
- FIELD_PREP(IRDMA_CQPSQ_UP_USEOVERRIDE,
- info->use_cnp_up_override) |
- FIELD_PREP(IRDMA_CQPSQ_UP_OP, IRDMA_CQP_OP_UP_MAP);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("WQE: UPMAP WQE", DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_manage_ws_node - create/modify/destroy WS node
- * @cqp: struct for cqp hw
- * @info: node info structure
- * @node_op: 0 for add 1 for modify, 2 for delete
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_manage_ws_node(struct irdma_sc_cqp *cqp,
- struct irdma_ws_node_info *info,
- enum irdma_ws_node_op node_op, u64 scratch)
- {
- __le64 *wqe;
- u64 temp = 0;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 32,
- FIELD_PREP(IRDMA_CQPSQ_WS_VSI, info->vsi) |
- FIELD_PREP(IRDMA_CQPSQ_WS_WEIGHT, info->weight));
- temp = FIELD_PREP(IRDMA_CQPSQ_WS_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_WS_NODEOP, node_op) |
- FIELD_PREP(IRDMA_CQPSQ_WS_ENABLENODE, info->enable) |
- FIELD_PREP(IRDMA_CQPSQ_WS_NODETYPE, info->type_leaf) |
- FIELD_PREP(IRDMA_CQPSQ_WS_PRIOTYPE, info->prio_type) |
- FIELD_PREP(IRDMA_CQPSQ_WS_TC, info->tc) |
- FIELD_PREP(IRDMA_CQPSQ_WS_OP, IRDMA_CQP_OP_WORK_SCHED_NODE) |
- FIELD_PREP(IRDMA_CQPSQ_WS_PARENTID, info->parent_id) |
- FIELD_PREP(IRDMA_CQPSQ_WS_NODEID, info->id);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("WQE: MANAGE_WS WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_qp_flush_wqes - flush qp's wqe
- * @qp: sc qp
- * @info: dlush information
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_qp_flush_wqes(struct irdma_sc_qp *qp,
- struct irdma_qp_flush_info *info, u64 scratch,
- bool post_sq)
- {
- u64 temp = 0;
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- bool flush_sq = false, flush_rq = false;
- if (info->rq && !qp->flush_rq)
- flush_rq = true;
- if (info->sq && !qp->flush_sq)
- flush_sq = true;
- qp->flush_sq |= flush_sq;
- qp->flush_rq |= flush_rq;
- if (!flush_sq && !flush_rq) {
- ibdev_dbg(to_ibdev(qp->dev),
- "CQP: Additional flush request ignored for qp %x\n",
- qp->qp_uk.qp_id);
- return -EALREADY;
- }
- cqp = qp->pd->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- if (info->userflushcode) {
- if (flush_rq)
- temp |= FIELD_PREP(IRDMA_CQPSQ_FWQE_RQMNERR,
- info->rq_minor_code) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_RQMJERR,
- info->rq_major_code);
- if (flush_sq)
- temp |= FIELD_PREP(IRDMA_CQPSQ_FWQE_SQMNERR,
- info->sq_minor_code) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_SQMJERR,
- info->sq_major_code);
- }
- set_64bit_val(wqe, 16, temp);
- temp = (info->generate_ae) ?
- info->ae_code | FIELD_PREP(IRDMA_CQPSQ_FWQE_AESOURCE,
- info->ae_src) : 0;
- set_64bit_val(wqe, 8, temp);
- hdr = qp->qp_uk.qp_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_FLUSH_WQES) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_GENERATE_AE, info->generate_ae) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_USERFLCODE, info->userflushcode) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_FLUSHSQ, flush_sq) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_FLUSHRQ, flush_rq) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QP_FLUSH WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_gen_ae - generate AE, uses flush WQE CQP OP
- * @qp: sc qp
- * @info: gen ae information
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_gen_ae(struct irdma_sc_qp *qp,
- struct irdma_gen_ae_info *info, u64 scratch,
- bool post_sq)
- {
- u64 temp;
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- cqp = qp->pd->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- temp = info->ae_code | FIELD_PREP(IRDMA_CQPSQ_FWQE_AESOURCE,
- info->ae_src);
- set_64bit_val(wqe, 8, temp);
- hdr = qp->qp_uk.qp_id | FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_GEN_AE) |
- FIELD_PREP(IRDMA_CQPSQ_FWQE_GENERATE_AE, 1) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: GEN_AE WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /*** irdma_sc_qp_upload_context - upload qp's context
- * @dev: sc device struct
- * @info: upload context info ptr for return
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_qp_upload_context(struct irdma_sc_dev *dev,
- struct irdma_upload_context_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, info->buf_pa);
- hdr = FIELD_PREP(IRDMA_CQPSQ_UCTX_QPID, info->qp_id) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_UPLOAD_CONTEXT) |
- FIELD_PREP(IRDMA_CQPSQ_UCTX_QPTYPE, info->qp_type) |
- FIELD_PREP(IRDMA_CQPSQ_UCTX_RAWFORMAT, info->raw_format) |
- FIELD_PREP(IRDMA_CQPSQ_UCTX_FREEZEQP, info->freeze_qp) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QP_UPLOAD_CTX WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_manage_push_page - Handle push page
- * @cqp: struct for cqp hw
- * @info: push page info
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_manage_push_page(struct irdma_sc_cqp *cqp,
- struct irdma_cqp_manage_push_page_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- if (info->free_page &&
- info->push_idx >= cqp->dev->hw_attrs.max_hw_device_pages)
- return -EINVAL;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, info->qs_handle);
- hdr = FIELD_PREP(IRDMA_CQPSQ_MPP_PPIDX, info->push_idx) |
- FIELD_PREP(IRDMA_CQPSQ_MPP_PPTYPE, info->push_page_type) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MANAGE_PUSH_PAGES) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_MPP_FREE_PAGE, info->free_page);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: MANAGE_PUSH_PAGES WQE", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_suspend_qp - suspend qp for param change
- * @cqp: struct for cqp hw
- * @qp: sc qp struct
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_suspend_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp,
- u64 scratch)
- {
- u64 hdr;
- __le64 *wqe;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- hdr = FIELD_PREP(IRDMA_CQPSQ_SUSPENDQP_QPID, qp->qp_uk.qp_id) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_SUSPEND_QP) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: SUSPEND_QP WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_resume_qp - resume qp after suspend
- * @cqp: struct for cqp hw
- * @qp: sc qp struct
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_resume_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp,
- u64 scratch)
- {
- u64 hdr;
- __le64 *wqe;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_RESUMEQP_QSHANDLE, qp->qs_handle));
- hdr = FIELD_PREP(IRDMA_CQPSQ_RESUMEQP_QPID, qp->qp_uk.qp_id) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_RESUME_QP) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: RESUME_QP WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_cq_ack - acknowledge completion q
- * @cq: cq struct
- */
- static inline void irdma_sc_cq_ack(struct irdma_sc_cq *cq)
- {
- writel(cq->cq_uk.cq_id, cq->cq_uk.cq_ack_db);
- }
- /**
- * irdma_sc_cq_init - initialize completion q
- * @cq: cq struct
- * @info: cq initialization info
- */
- int irdma_sc_cq_init(struct irdma_sc_cq *cq, struct irdma_cq_init_info *info)
- {
- u32 pble_obj_cnt;
- pble_obj_cnt = info->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->virtual_map && info->first_pm_pbl_idx >= pble_obj_cnt)
- return -EINVAL;
- cq->cq_pa = info->cq_base_pa;
- cq->dev = info->dev;
- cq->ceq_id = info->ceq_id;
- info->cq_uk_init_info.cqe_alloc_db = cq->dev->cq_arm_db;
- info->cq_uk_init_info.cq_ack_db = cq->dev->cq_ack_db;
- irdma_uk_cq_init(&cq->cq_uk, &info->cq_uk_init_info);
- cq->virtual_map = info->virtual_map;
- cq->pbl_chunk_size = info->pbl_chunk_size;
- cq->ceqe_mask = info->ceqe_mask;
- cq->cq_type = (info->type) ? info->type : IRDMA_CQ_TYPE_IWARP;
- cq->shadow_area_pa = info->shadow_area_pa;
- cq->shadow_read_threshold = info->shadow_read_threshold;
- cq->ceq_id_valid = info->ceq_id_valid;
- cq->tph_en = info->tph_en;
- cq->tph_val = info->tph_val;
- cq->first_pm_pbl_idx = info->first_pm_pbl_idx;
- cq->vsi = info->vsi;
- return 0;
- }
- /**
- * irdma_sc_cq_create - create completion q
- * @cq: cq struct
- * @scratch: u64 saved to be used during cqp completion
- * @check_overflow: flag for overflow check
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_cq_create(struct irdma_sc_cq *cq, u64 scratch,
- bool check_overflow, bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- struct irdma_sc_ceq *ceq;
- int ret_code = 0;
- cqp = cq->dev->cqp;
- if (cq->cq_uk.cq_id >= cqp->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].max_cnt)
- return -EINVAL;
- if (cq->ceq_id >= cq->dev->hmc_fpm_misc.max_ceqs)
- return -EINVAL;
- ceq = cq->dev->ceq[cq->ceq_id];
- if (ceq && ceq->reg_cq)
- ret_code = irdma_sc_add_cq_ctx(ceq, cq);
- if (ret_code)
- return ret_code;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe) {
- if (ceq && ceq->reg_cq)
- irdma_sc_remove_cq_ctx(ceq, cq);
- return -ENOMEM;
- }
- set_64bit_val(wqe, 0, cq->cq_uk.cq_size);
- set_64bit_val(wqe, 8, (uintptr_t)cq >> 1);
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_CQ_SHADOW_READ_THRESHOLD, cq->shadow_read_threshold));
- set_64bit_val(wqe, 32, (cq->virtual_map ? 0 : cq->cq_pa));
- set_64bit_val(wqe, 40, cq->shadow_area_pa);
- set_64bit_val(wqe, 48,
- FIELD_PREP(IRDMA_CQPSQ_CQ_FIRSTPMPBLIDX, (cq->virtual_map ? cq->first_pm_pbl_idx : 0)));
- set_64bit_val(wqe, 56,
- FIELD_PREP(IRDMA_CQPSQ_TPHVAL, cq->tph_val) |
- FIELD_PREP(IRDMA_CQPSQ_VSIIDX, cq->vsi->vsi_idx));
- hdr = FLD_LS_64(cq->dev, cq->cq_uk.cq_id, IRDMA_CQPSQ_CQ_CQID) |
- FLD_LS_64(cq->dev, (cq->ceq_id_valid ? cq->ceq_id : 0),
- IRDMA_CQPSQ_CQ_CEQID) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_CREATE_CQ) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_LPBLSIZE, cq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CHKOVERFLOW, check_overflow) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_VIRTMAP, cq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_ENCEQEMASK, cq->ceqe_mask) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CEQIDVALID, cq->ceq_id_valid) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, cq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_AVOIDMEMCNFLCT,
- cq->cq_uk.avoid_mem_cflct) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CQ_CREATE WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_cq_destroy - destroy completion q
- * @cq: cq struct
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_cq_destroy(struct irdma_sc_cq *cq, u64 scratch, bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- struct irdma_sc_ceq *ceq;
- cqp = cq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- ceq = cq->dev->ceq[cq->ceq_id];
- if (ceq && ceq->reg_cq)
- irdma_sc_remove_cq_ctx(ceq, cq);
- set_64bit_val(wqe, 0, cq->cq_uk.cq_size);
- set_64bit_val(wqe, 8, (uintptr_t)cq >> 1);
- set_64bit_val(wqe, 40, cq->shadow_area_pa);
- set_64bit_val(wqe, 48,
- (cq->virtual_map ? cq->first_pm_pbl_idx : 0));
- hdr = cq->cq_uk.cq_id |
- FLD_LS_64(cq->dev, (cq->ceq_id_valid ? cq->ceq_id : 0),
- IRDMA_CQPSQ_CQ_CEQID) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DESTROY_CQ) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_LPBLSIZE, cq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_VIRTMAP, cq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_ENCEQEMASK, cq->ceqe_mask) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CEQIDVALID, cq->ceq_id_valid) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, cq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_AVOIDMEMCNFLCT, cq->cq_uk.avoid_mem_cflct) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CQ_DESTROY WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_cq_resize - set resized cq buffer info
- * @cq: resized cq
- * @info: resized cq buffer info
- */
- void irdma_sc_cq_resize(struct irdma_sc_cq *cq, struct irdma_modify_cq_info *info)
- {
- cq->virtual_map = info->virtual_map;
- cq->cq_pa = info->cq_pa;
- cq->first_pm_pbl_idx = info->first_pm_pbl_idx;
- cq->pbl_chunk_size = info->pbl_chunk_size;
- irdma_uk_cq_resize(&cq->cq_uk, info->cq_base, info->cq_size);
- }
- /**
- * irdma_sc_cq_modify - modify a Completion Queue
- * @cq: cq struct
- * @info: modification info struct
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag to post to sq
- */
- static int irdma_sc_cq_modify(struct irdma_sc_cq *cq,
- struct irdma_modify_cq_info *info, u64 scratch,
- bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- u32 pble_obj_cnt;
- pble_obj_cnt = cq->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->cq_resize && info->virtual_map &&
- info->first_pm_pbl_idx >= pble_obj_cnt)
- return -EINVAL;
- cqp = cq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 0, info->cq_size);
- set_64bit_val(wqe, 8, (uintptr_t)cq >> 1);
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_CQPSQ_CQ_SHADOW_READ_THRESHOLD, info->shadow_read_threshold));
- set_64bit_val(wqe, 32, info->cq_pa);
- set_64bit_val(wqe, 40, cq->shadow_area_pa);
- set_64bit_val(wqe, 48, info->first_pm_pbl_idx);
- set_64bit_val(wqe, 56,
- FIELD_PREP(IRDMA_CQPSQ_TPHVAL, cq->tph_val) |
- FIELD_PREP(IRDMA_CQPSQ_VSIIDX, cq->vsi->vsi_idx));
- hdr = cq->cq_uk.cq_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_MODIFY_CQ) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CQRESIZE, info->cq_resize) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_LPBLSIZE, info->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CHKOVERFLOW, info->check_overflow) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_VIRTMAP, info->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_ENCEQEMASK, cq->ceqe_mask) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, cq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_AVOIDMEMCNFLCT,
- cq->cq_uk.avoid_mem_cflct) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CQ_MODIFY WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_check_cqp_progress - check cqp processing progress
- * @timeout: timeout info struct
- * @dev: sc device struct
- */
- void irdma_check_cqp_progress(struct irdma_cqp_timeout *timeout, struct irdma_sc_dev *dev)
- {
- u64 completed_ops = atomic64_read(&dev->cqp->completed_ops);
- if (timeout->compl_cqp_cmds != completed_ops) {
- timeout->compl_cqp_cmds = completed_ops;
- timeout->count = 0;
- } else if (timeout->compl_cqp_cmds != dev->cqp->requested_ops) {
- timeout->count++;
- }
- }
- /**
- * irdma_get_cqp_reg_info - get head and tail for cqp using registers
- * @cqp: struct for cqp hw
- * @val: cqp tail register value
- * @tail: wqtail register value
- * @error: cqp processing err
- */
- static inline void irdma_get_cqp_reg_info(struct irdma_sc_cqp *cqp, u32 *val,
- u32 *tail, u32 *error)
- {
- *val = readl(cqp->dev->hw_regs[IRDMA_CQPTAIL]);
- *tail = FIELD_GET(IRDMA_CQPTAIL_WQTAIL, *val);
- *error = FIELD_GET(IRDMA_CQPTAIL_CQP_OP_ERR, *val);
- }
- /**
- * irdma_cqp_poll_registers - poll cqp registers
- * @cqp: struct for cqp hw
- * @tail: wqtail register value
- * @count: how many times to try for completion
- */
- static int irdma_cqp_poll_registers(struct irdma_sc_cqp *cqp, u32 tail,
- u32 count)
- {
- u32 i = 0;
- u32 newtail, error, val;
- while (i++ < count) {
- irdma_get_cqp_reg_info(cqp, &val, &newtail, &error);
- if (error) {
- error = readl(cqp->dev->hw_regs[IRDMA_CQPERRCODES]);
- ibdev_dbg(to_ibdev(cqp->dev),
- "CQP: CQPERRCODES error_code[x%08X]\n",
- error);
- return -EIO;
- }
- if (newtail != tail) {
- /* SUCCESS */
- IRDMA_RING_MOVE_TAIL(cqp->sq_ring);
- atomic64_inc(&cqp->completed_ops);
- return 0;
- }
- udelay(cqp->dev->hw_attrs.max_sleep_count);
- }
- return -ETIMEDOUT;
- }
- /**
- * irdma_sc_decode_fpm_commit - decode a 64 bit value into count and base
- * @dev: sc device struct
- * @buf: pointer to commit buffer
- * @buf_idx: buffer index
- * @obj_info: object info pointer
- * @rsrc_idx: indexs of memory resource
- */
- static u64 irdma_sc_decode_fpm_commit(struct irdma_sc_dev *dev, __le64 *buf,
- u32 buf_idx, struct irdma_hmc_obj_info *obj_info,
- u32 rsrc_idx)
- {
- u64 temp;
- get_64bit_val(buf, buf_idx, &temp);
- switch (rsrc_idx) {
- case IRDMA_HMC_IW_QP:
- obj_info[rsrc_idx].cnt = (u32)FIELD_GET(IRDMA_COMMIT_FPM_QPCNT, temp);
- break;
- case IRDMA_HMC_IW_CQ:
- obj_info[rsrc_idx].cnt = (u32)FLD_RS_64(dev, temp, IRDMA_COMMIT_FPM_CQCNT);
- break;
- case IRDMA_HMC_IW_APBVT_ENTRY:
- obj_info[rsrc_idx].cnt = 1;
- break;
- default:
- obj_info[rsrc_idx].cnt = (u32)temp;
- break;
- }
- obj_info[rsrc_idx].base = (temp >> IRDMA_COMMIT_FPM_BASE_S) * 512;
- return temp;
- }
- /**
- * irdma_sc_parse_fpm_commit_buf - parse fpm commit buffer
- * @dev: pointer to dev struct
- * @buf: ptr to fpm commit buffer
- * @info: ptr to irdma_hmc_obj_info struct
- * @sd: number of SDs for HMC objects
- *
- * parses fpm commit info and copy base value
- * of hmc objects in hmc_info
- */
- static void
- irdma_sc_parse_fpm_commit_buf(struct irdma_sc_dev *dev, __le64 *buf,
- struct irdma_hmc_obj_info *info, u32 *sd)
- {
- u64 size;
- u32 i;
- u64 max_base = 0;
- u32 last_hmc_obj = 0;
- irdma_sc_decode_fpm_commit(dev, buf, 0, info,
- IRDMA_HMC_IW_QP);
- irdma_sc_decode_fpm_commit(dev, buf, 8, info,
- IRDMA_HMC_IW_CQ);
- /* skiping RSRVD */
- irdma_sc_decode_fpm_commit(dev, buf, 24, info,
- IRDMA_HMC_IW_HTE);
- irdma_sc_decode_fpm_commit(dev, buf, 32, info,
- IRDMA_HMC_IW_ARP);
- irdma_sc_decode_fpm_commit(dev, buf, 40, info,
- IRDMA_HMC_IW_APBVT_ENTRY);
- irdma_sc_decode_fpm_commit(dev, buf, 48, info,
- IRDMA_HMC_IW_MR);
- irdma_sc_decode_fpm_commit(dev, buf, 56, info,
- IRDMA_HMC_IW_XF);
- irdma_sc_decode_fpm_commit(dev, buf, 64, info,
- IRDMA_HMC_IW_XFFL);
- irdma_sc_decode_fpm_commit(dev, buf, 72, info,
- IRDMA_HMC_IW_Q1);
- irdma_sc_decode_fpm_commit(dev, buf, 80, info,
- IRDMA_HMC_IW_Q1FL);
- irdma_sc_decode_fpm_commit(dev, buf, 88, info,
- IRDMA_HMC_IW_TIMER);
- irdma_sc_decode_fpm_commit(dev, buf, 112, info,
- IRDMA_HMC_IW_PBLE);
- /* skipping RSVD. */
- if (dev->hw_attrs.uk_attrs.hw_rev != IRDMA_GEN_1) {
- irdma_sc_decode_fpm_commit(dev, buf, 96, info,
- IRDMA_HMC_IW_FSIMC);
- irdma_sc_decode_fpm_commit(dev, buf, 104, info,
- IRDMA_HMC_IW_FSIAV);
- irdma_sc_decode_fpm_commit(dev, buf, 128, info,
- IRDMA_HMC_IW_RRF);
- irdma_sc_decode_fpm_commit(dev, buf, 136, info,
- IRDMA_HMC_IW_RRFFL);
- irdma_sc_decode_fpm_commit(dev, buf, 144, info,
- IRDMA_HMC_IW_HDR);
- irdma_sc_decode_fpm_commit(dev, buf, 152, info,
- IRDMA_HMC_IW_MD);
- irdma_sc_decode_fpm_commit(dev, buf, 160, info,
- IRDMA_HMC_IW_OOISC);
- irdma_sc_decode_fpm_commit(dev, buf, 168, info,
- IRDMA_HMC_IW_OOISCFFL);
- }
- /* searching for the last object in HMC to find the size of the HMC area. */
- for (i = IRDMA_HMC_IW_QP; i < IRDMA_HMC_IW_MAX; i++) {
- if (info[i].base > max_base) {
- max_base = info[i].base;
- last_hmc_obj = i;
- }
- }
- size = info[last_hmc_obj].cnt * info[last_hmc_obj].size +
- info[last_hmc_obj].base;
- if (size & 0x1FFFFF)
- *sd = (u32)((size >> 21) + 1); /* add 1 for remainder */
- else
- *sd = (u32)(size >> 21);
- }
- /**
- * irdma_sc_decode_fpm_query() - Decode a 64 bit value into max count and size
- * @buf: ptr to fpm query buffer
- * @buf_idx: index into buf
- * @obj_info: ptr to irdma_hmc_obj_info struct
- * @rsrc_idx: resource index into info
- *
- * Decode a 64 bit value from fpm query buffer into max count and size
- */
- static u64 irdma_sc_decode_fpm_query(__le64 *buf, u32 buf_idx,
- struct irdma_hmc_obj_info *obj_info,
- u32 rsrc_idx)
- {
- u64 temp;
- u32 size;
- get_64bit_val(buf, buf_idx, &temp);
- obj_info[rsrc_idx].max_cnt = (u32)temp;
- size = (u32)(temp >> 32);
- obj_info[rsrc_idx].size = BIT_ULL(size);
- return temp;
- }
- /**
- * irdma_sc_parse_fpm_query_buf() - parses fpm query buffer
- * @dev: ptr to shared code device
- * @buf: ptr to fpm query buffer
- * @hmc_info: ptr to irdma_hmc_obj_info struct
- * @hmc_fpm_misc: ptr to fpm data
- *
- * parses fpm query buffer and copy max_cnt and
- * size value of hmc objects in hmc_info
- */
- static int irdma_sc_parse_fpm_query_buf(struct irdma_sc_dev *dev, __le64 *buf,
- struct irdma_hmc_info *hmc_info,
- struct irdma_hmc_fpm_misc *hmc_fpm_misc)
- {
- struct irdma_hmc_obj_info *obj_info;
- u64 temp;
- u32 size;
- u16 max_pe_sds;
- obj_info = hmc_info->hmc_obj;
- get_64bit_val(buf, 0, &temp);
- hmc_info->first_sd_index = (u16)FIELD_GET(IRDMA_QUERY_FPM_FIRST_PE_SD_INDEX, temp);
- max_pe_sds = (u16)FIELD_GET(IRDMA_QUERY_FPM_MAX_PE_SDS, temp);
- hmc_fpm_misc->max_sds = max_pe_sds;
- hmc_info->sd_table.sd_cnt = max_pe_sds + hmc_info->first_sd_index;
- get_64bit_val(buf, 8, &temp);
- obj_info[IRDMA_HMC_IW_QP].max_cnt = (u32)FIELD_GET(IRDMA_QUERY_FPM_MAX_QPS, temp);
- size = (u32)(temp >> 32);
- obj_info[IRDMA_HMC_IW_QP].size = BIT_ULL(size);
- get_64bit_val(buf, 16, &temp);
- obj_info[IRDMA_HMC_IW_CQ].max_cnt = (u32)FIELD_GET(IRDMA_QUERY_FPM_MAX_CQS, temp);
- size = (u32)(temp >> 32);
- obj_info[IRDMA_HMC_IW_CQ].size = BIT_ULL(size);
- irdma_sc_decode_fpm_query(buf, 32, obj_info, IRDMA_HMC_IW_HTE);
- irdma_sc_decode_fpm_query(buf, 40, obj_info, IRDMA_HMC_IW_ARP);
- obj_info[IRDMA_HMC_IW_APBVT_ENTRY].size = 8192;
- obj_info[IRDMA_HMC_IW_APBVT_ENTRY].max_cnt = 1;
- irdma_sc_decode_fpm_query(buf, 48, obj_info, IRDMA_HMC_IW_MR);
- irdma_sc_decode_fpm_query(buf, 56, obj_info, IRDMA_HMC_IW_XF);
- get_64bit_val(buf, 64, &temp);
- obj_info[IRDMA_HMC_IW_XFFL].max_cnt = (u32)temp;
- obj_info[IRDMA_HMC_IW_XFFL].size = 4;
- hmc_fpm_misc->xf_block_size = FIELD_GET(IRDMA_QUERY_FPM_XFBLOCKSIZE, temp);
- if (!hmc_fpm_misc->xf_block_size)
- return -EINVAL;
- irdma_sc_decode_fpm_query(buf, 72, obj_info, IRDMA_HMC_IW_Q1);
- get_64bit_val(buf, 80, &temp);
- obj_info[IRDMA_HMC_IW_Q1FL].max_cnt = (u32)temp;
- obj_info[IRDMA_HMC_IW_Q1FL].size = 4;
- hmc_fpm_misc->q1_block_size = FIELD_GET(IRDMA_QUERY_FPM_Q1BLOCKSIZE, temp);
- if (!hmc_fpm_misc->q1_block_size)
- return -EINVAL;
- irdma_sc_decode_fpm_query(buf, 88, obj_info, IRDMA_HMC_IW_TIMER);
- get_64bit_val(buf, 112, &temp);
- obj_info[IRDMA_HMC_IW_PBLE].max_cnt = (u32)temp;
- obj_info[IRDMA_HMC_IW_PBLE].size = 8;
- get_64bit_val(buf, 120, &temp);
- hmc_fpm_misc->max_ceqs = FIELD_GET(IRDMA_QUERY_FPM_MAX_CEQS, temp);
- hmc_fpm_misc->ht_multiplier = FIELD_GET(IRDMA_QUERY_FPM_HTMULTIPLIER, temp);
- hmc_fpm_misc->timer_bucket = FIELD_GET(IRDMA_QUERY_FPM_TIMERBUCKET, temp);
- if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1)
- return 0;
- irdma_sc_decode_fpm_query(buf, 96, obj_info, IRDMA_HMC_IW_FSIMC);
- irdma_sc_decode_fpm_query(buf, 104, obj_info, IRDMA_HMC_IW_FSIAV);
- irdma_sc_decode_fpm_query(buf, 128, obj_info, IRDMA_HMC_IW_RRF);
- get_64bit_val(buf, 136, &temp);
- obj_info[IRDMA_HMC_IW_RRFFL].max_cnt = (u32)temp;
- obj_info[IRDMA_HMC_IW_RRFFL].size = 4;
- hmc_fpm_misc->rrf_block_size = FIELD_GET(IRDMA_QUERY_FPM_RRFBLOCKSIZE, temp);
- if (!hmc_fpm_misc->rrf_block_size &&
- obj_info[IRDMA_HMC_IW_RRFFL].max_cnt)
- return -EINVAL;
- irdma_sc_decode_fpm_query(buf, 144, obj_info, IRDMA_HMC_IW_HDR);
- irdma_sc_decode_fpm_query(buf, 152, obj_info, IRDMA_HMC_IW_MD);
- irdma_sc_decode_fpm_query(buf, 160, obj_info, IRDMA_HMC_IW_OOISC);
- get_64bit_val(buf, 168, &temp);
- obj_info[IRDMA_HMC_IW_OOISCFFL].max_cnt = (u32)temp;
- obj_info[IRDMA_HMC_IW_OOISCFFL].size = 4;
- hmc_fpm_misc->ooiscf_block_size = FIELD_GET(IRDMA_QUERY_FPM_OOISCFBLOCKSIZE, temp);
- if (!hmc_fpm_misc->ooiscf_block_size &&
- obj_info[IRDMA_HMC_IW_OOISCFFL].max_cnt)
- return -EINVAL;
- return 0;
- }
- /**
- * irdma_sc_find_reg_cq - find cq ctx index
- * @ceq: ceq sc structure
- * @cq: cq sc structure
- */
- static u32 irdma_sc_find_reg_cq(struct irdma_sc_ceq *ceq,
- struct irdma_sc_cq *cq)
- {
- u32 i;
- for (i = 0; i < ceq->reg_cq_size; i++) {
- if (cq == ceq->reg_cq[i])
- return i;
- }
- return IRDMA_INVALID_CQ_IDX;
- }
- /**
- * irdma_sc_add_cq_ctx - add cq ctx tracking for ceq
- * @ceq: ceq sc structure
- * @cq: cq sc structure
- */
- int irdma_sc_add_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq)
- {
- unsigned long flags;
- spin_lock_irqsave(&ceq->req_cq_lock, flags);
- if (ceq->reg_cq_size == ceq->elem_cnt) {
- spin_unlock_irqrestore(&ceq->req_cq_lock, flags);
- return -ENOMEM;
- }
- ceq->reg_cq[ceq->reg_cq_size++] = cq;
- spin_unlock_irqrestore(&ceq->req_cq_lock, flags);
- return 0;
- }
- /**
- * irdma_sc_remove_cq_ctx - remove cq ctx tracking for ceq
- * @ceq: ceq sc structure
- * @cq: cq sc structure
- */
- void irdma_sc_remove_cq_ctx(struct irdma_sc_ceq *ceq, struct irdma_sc_cq *cq)
- {
- unsigned long flags;
- u32 cq_ctx_idx;
- spin_lock_irqsave(&ceq->req_cq_lock, flags);
- cq_ctx_idx = irdma_sc_find_reg_cq(ceq, cq);
- if (cq_ctx_idx == IRDMA_INVALID_CQ_IDX)
- goto exit;
- ceq->reg_cq_size--;
- if (cq_ctx_idx != ceq->reg_cq_size)
- ceq->reg_cq[cq_ctx_idx] = ceq->reg_cq[ceq->reg_cq_size];
- ceq->reg_cq[ceq->reg_cq_size] = NULL;
- exit:
- spin_unlock_irqrestore(&ceq->req_cq_lock, flags);
- }
- /**
- * irdma_sc_cqp_init - Initialize buffers for a control Queue Pair
- * @cqp: IWARP control queue pair pointer
- * @info: IWARP control queue pair init info pointer
- *
- * Initializes the object and context buffers for a control Queue Pair.
- */
- int irdma_sc_cqp_init(struct irdma_sc_cqp *cqp,
- struct irdma_cqp_init_info *info)
- {
- u8 hw_sq_size;
- if (info->sq_size > IRDMA_CQP_SW_SQSIZE_2048 ||
- info->sq_size < IRDMA_CQP_SW_SQSIZE_4 ||
- ((info->sq_size & (info->sq_size - 1))))
- return -EINVAL;
- hw_sq_size = irdma_get_encoded_wqe_size(info->sq_size,
- IRDMA_QUEUE_TYPE_CQP);
- cqp->size = sizeof(*cqp);
- cqp->sq_size = info->sq_size;
- cqp->hw_sq_size = hw_sq_size;
- cqp->sq_base = info->sq;
- cqp->host_ctx = info->host_ctx;
- cqp->sq_pa = info->sq_pa;
- cqp->host_ctx_pa = info->host_ctx_pa;
- cqp->dev = info->dev;
- cqp->struct_ver = info->struct_ver;
- cqp->hw_maj_ver = info->hw_maj_ver;
- cqp->hw_min_ver = info->hw_min_ver;
- cqp->scratch_array = info->scratch_array;
- cqp->polarity = 0;
- cqp->en_datacenter_tcp = info->en_datacenter_tcp;
- cqp->ena_vf_count = info->ena_vf_count;
- cqp->hmc_profile = info->hmc_profile;
- cqp->ceqs_per_vf = info->ceqs_per_vf;
- cqp->disable_packed = info->disable_packed;
- cqp->rocev2_rto_policy = info->rocev2_rto_policy;
- cqp->protocol_used = info->protocol_used;
- memcpy(&cqp->dcqcn_params, &info->dcqcn_params, sizeof(cqp->dcqcn_params));
- info->dev->cqp = cqp;
- IRDMA_RING_INIT(cqp->sq_ring, cqp->sq_size);
- cqp->requested_ops = 0;
- atomic64_set(&cqp->completed_ops, 0);
- /* for the cqp commands backlog. */
- INIT_LIST_HEAD(&cqp->dev->cqp_cmd_head);
- writel(0, cqp->dev->hw_regs[IRDMA_CQPTAIL]);
- writel(0, cqp->dev->hw_regs[IRDMA_CQPDB]);
- writel(0, cqp->dev->hw_regs[IRDMA_CCQPSTATUS]);
- ibdev_dbg(to_ibdev(cqp->dev),
- "WQE: sq_size[%04d] hw_sq_size[%04d] sq_base[%p] sq_pa[%pK] cqp[%p] polarity[x%04x]\n",
- cqp->sq_size, cqp->hw_sq_size, cqp->sq_base,
- (u64 *)(uintptr_t)cqp->sq_pa, cqp, cqp->polarity);
- return 0;
- }
- /**
- * irdma_sc_cqp_create - create cqp during bringup
- * @cqp: struct for cqp hw
- * @maj_err: If error, major err number
- * @min_err: If error, minor err number
- */
- int irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err)
- {
- u64 temp;
- u8 hw_rev;
- u32 cnt = 0, p1, p2, val = 0, err_code;
- int ret_code;
- hw_rev = cqp->dev->hw_attrs.uk_attrs.hw_rev;
- cqp->sdbuf.size = ALIGN(IRDMA_UPDATE_SD_BUFF_SIZE * cqp->sq_size,
- IRDMA_SD_BUF_ALIGNMENT);
- cqp->sdbuf.va = dma_alloc_coherent(cqp->dev->hw->device,
- cqp->sdbuf.size, &cqp->sdbuf.pa,
- GFP_KERNEL);
- if (!cqp->sdbuf.va)
- return -ENOMEM;
- spin_lock_init(&cqp->dev->cqp_lock);
- temp = FIELD_PREP(IRDMA_CQPHC_SQSIZE, cqp->hw_sq_size) |
- FIELD_PREP(IRDMA_CQPHC_SVER, cqp->struct_ver) |
- FIELD_PREP(IRDMA_CQPHC_DISABLE_PFPDUS, cqp->disable_packed) |
- FIELD_PREP(IRDMA_CQPHC_CEQPERVF, cqp->ceqs_per_vf);
- if (hw_rev >= IRDMA_GEN_2) {
- temp |= FIELD_PREP(IRDMA_CQPHC_ROCEV2_RTO_POLICY,
- cqp->rocev2_rto_policy) |
- FIELD_PREP(IRDMA_CQPHC_PROTOCOL_USED,
- cqp->protocol_used);
- }
- set_64bit_val(cqp->host_ctx, 0, temp);
- set_64bit_val(cqp->host_ctx, 8, cqp->sq_pa);
- temp = FIELD_PREP(IRDMA_CQPHC_ENABLED_VFS, cqp->ena_vf_count) |
- FIELD_PREP(IRDMA_CQPHC_HMC_PROFILE, cqp->hmc_profile);
- set_64bit_val(cqp->host_ctx, 16, temp);
- set_64bit_val(cqp->host_ctx, 24, (uintptr_t)cqp);
- temp = FIELD_PREP(IRDMA_CQPHC_HW_MAJVER, cqp->hw_maj_ver) |
- FIELD_PREP(IRDMA_CQPHC_HW_MINVER, cqp->hw_min_ver);
- if (hw_rev >= IRDMA_GEN_2) {
- temp |= FIELD_PREP(IRDMA_CQPHC_MIN_RATE, cqp->dcqcn_params.min_rate) |
- FIELD_PREP(IRDMA_CQPHC_MIN_DEC_FACTOR, cqp->dcqcn_params.min_dec_factor);
- }
- set_64bit_val(cqp->host_ctx, 32, temp);
- set_64bit_val(cqp->host_ctx, 40, 0);
- temp = 0;
- if (hw_rev >= IRDMA_GEN_2) {
- temp |= FIELD_PREP(IRDMA_CQPHC_DCQCN_T, cqp->dcqcn_params.dcqcn_t) |
- FIELD_PREP(IRDMA_CQPHC_RAI_FACTOR, cqp->dcqcn_params.rai_factor) |
- FIELD_PREP(IRDMA_CQPHC_HAI_FACTOR, cqp->dcqcn_params.hai_factor);
- }
- set_64bit_val(cqp->host_ctx, 48, temp);
- temp = 0;
- if (hw_rev >= IRDMA_GEN_2) {
- temp |= FIELD_PREP(IRDMA_CQPHC_DCQCN_B, cqp->dcqcn_params.dcqcn_b) |
- FIELD_PREP(IRDMA_CQPHC_DCQCN_F, cqp->dcqcn_params.dcqcn_f) |
- FIELD_PREP(IRDMA_CQPHC_CC_CFG_VALID, cqp->dcqcn_params.cc_cfg_valid) |
- FIELD_PREP(IRDMA_CQPHC_RREDUCE_MPERIOD, cqp->dcqcn_params.rreduce_mperiod);
- }
- set_64bit_val(cqp->host_ctx, 56, temp);
- print_hex_dump_debug("WQE: CQP_HOST_CTX WQE", DUMP_PREFIX_OFFSET, 16,
- 8, cqp->host_ctx, IRDMA_CQP_CTX_SIZE * 8, false);
- p1 = cqp->host_ctx_pa >> 32;
- p2 = (u32)cqp->host_ctx_pa;
- writel(p1, cqp->dev->hw_regs[IRDMA_CCQPHIGH]);
- writel(p2, cqp->dev->hw_regs[IRDMA_CCQPLOW]);
- do {
- if (cnt++ > cqp->dev->hw_attrs.max_done_count) {
- ret_code = -ETIMEDOUT;
- goto err;
- }
- udelay(cqp->dev->hw_attrs.max_sleep_count);
- val = readl(cqp->dev->hw_regs[IRDMA_CCQPSTATUS]);
- } while (!val);
- if (FLD_RS_32(cqp->dev, val, IRDMA_CCQPSTATUS_CCQP_ERR)) {
- ret_code = -EOPNOTSUPP;
- goto err;
- }
- cqp->process_cqp_sds = irdma_update_sds_noccq;
- return 0;
- err:
- dma_free_coherent(cqp->dev->hw->device, cqp->sdbuf.size,
- cqp->sdbuf.va, cqp->sdbuf.pa);
- cqp->sdbuf.va = NULL;
- err_code = readl(cqp->dev->hw_regs[IRDMA_CQPERRCODES]);
- *min_err = FIELD_GET(IRDMA_CQPERRCODES_CQP_MINOR_CODE, err_code);
- *maj_err = FIELD_GET(IRDMA_CQPERRCODES_CQP_MAJOR_CODE, err_code);
- return ret_code;
- }
- /**
- * irdma_sc_cqp_post_sq - post of cqp's sq
- * @cqp: struct for cqp hw
- */
- void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp)
- {
- writel(IRDMA_RING_CURRENT_HEAD(cqp->sq_ring), cqp->dev->cqp_db);
- ibdev_dbg(to_ibdev(cqp->dev),
- "WQE: CQP SQ head 0x%x tail 0x%x size 0x%x\n",
- cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size);
- }
- /**
- * irdma_sc_cqp_get_next_send_wqe_idx - get next wqe on cqp sq
- * and pass back index
- * @cqp: CQP HW structure
- * @scratch: private data for CQP WQE
- * @wqe_idx: WQE index of CQP SQ
- */
- __le64 *irdma_sc_cqp_get_next_send_wqe_idx(struct irdma_sc_cqp *cqp, u64 scratch,
- u32 *wqe_idx)
- {
- __le64 *wqe = NULL;
- int ret_code;
- if (IRDMA_RING_FULL_ERR(cqp->sq_ring)) {
- ibdev_dbg(to_ibdev(cqp->dev),
- "WQE: CQP SQ is full, head 0x%x tail 0x%x size 0x%x\n",
- cqp->sq_ring.head, cqp->sq_ring.tail,
- cqp->sq_ring.size);
- return NULL;
- }
- IRDMA_ATOMIC_RING_MOVE_HEAD(cqp->sq_ring, *wqe_idx, ret_code);
- if (ret_code)
- return NULL;
- cqp->requested_ops++;
- if (!*wqe_idx)
- cqp->polarity = !cqp->polarity;
- wqe = cqp->sq_base[*wqe_idx].elem;
- cqp->scratch_array[*wqe_idx] = scratch;
- IRDMA_CQP_INIT_WQE(wqe);
- return wqe;
- }
- /**
- * irdma_sc_cqp_destroy - destroy cqp during close
- * @cqp: struct for cqp hw
- */
- int irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp)
- {
- u32 cnt = 0, val;
- int ret_code = 0;
- writel(0, cqp->dev->hw_regs[IRDMA_CCQPHIGH]);
- writel(0, cqp->dev->hw_regs[IRDMA_CCQPLOW]);
- do {
- if (cnt++ > cqp->dev->hw_attrs.max_done_count) {
- ret_code = -ETIMEDOUT;
- break;
- }
- udelay(cqp->dev->hw_attrs.max_sleep_count);
- val = readl(cqp->dev->hw_regs[IRDMA_CCQPSTATUS]);
- } while (FLD_RS_32(cqp->dev, val, IRDMA_CCQPSTATUS_CCQP_DONE));
- dma_free_coherent(cqp->dev->hw->device, cqp->sdbuf.size,
- cqp->sdbuf.va, cqp->sdbuf.pa);
- cqp->sdbuf.va = NULL;
- return ret_code;
- }
- /**
- * irdma_sc_ccq_arm - enable intr for control cq
- * @ccq: ccq sc struct
- */
- void irdma_sc_ccq_arm(struct irdma_sc_cq *ccq)
- {
- u64 temp_val;
- u16 sw_cq_sel;
- u8 arm_next_se;
- u8 arm_seq_num;
- get_64bit_val(ccq->cq_uk.shadow_area, 32, &temp_val);
- sw_cq_sel = (u16)FIELD_GET(IRDMA_CQ_DBSA_SW_CQ_SELECT, temp_val);
- arm_next_se = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_NEXT_SE, temp_val);
- arm_seq_num = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_SEQ_NUM, temp_val);
- arm_seq_num++;
- temp_val = FIELD_PREP(IRDMA_CQ_DBSA_ARM_SEQ_NUM, arm_seq_num) |
- FIELD_PREP(IRDMA_CQ_DBSA_SW_CQ_SELECT, sw_cq_sel) |
- FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT_SE, arm_next_se) |
- FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT, 1);
- set_64bit_val(ccq->cq_uk.shadow_area, 32, temp_val);
- dma_wmb(); /* make sure shadow area is updated before arming */
- writel(ccq->cq_uk.cq_id, ccq->dev->cq_arm_db);
- }
- /**
- * irdma_sc_ccq_get_cqe_info - get ccq's cq entry
- * @ccq: ccq sc struct
- * @info: completion q entry to return
- */
- int irdma_sc_ccq_get_cqe_info(struct irdma_sc_cq *ccq,
- struct irdma_ccq_cqe_info *info)
- {
- u64 qp_ctx, temp, temp1;
- __le64 *cqe;
- struct irdma_sc_cqp *cqp;
- u32 wqe_idx;
- u32 error;
- u8 polarity;
- int ret_code = 0;
- if (ccq->cq_uk.avoid_mem_cflct)
- cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(&ccq->cq_uk);
- else
- cqe = IRDMA_GET_CURRENT_CQ_ELEM(&ccq->cq_uk);
- get_64bit_val(cqe, 24, &temp);
- polarity = (u8)FIELD_GET(IRDMA_CQ_VALID, temp);
- if (polarity != ccq->cq_uk.polarity)
- return -ENOENT;
- /* Ensure CEQE contents are read after valid bit is checked */
- dma_rmb();
- get_64bit_val(cqe, 8, &qp_ctx);
- cqp = (struct irdma_sc_cqp *)(unsigned long)qp_ctx;
- info->error = (bool)FIELD_GET(IRDMA_CQ_ERROR, temp);
- info->maj_err_code = IRDMA_CQPSQ_MAJ_NO_ERROR;
- info->min_err_code = (u16)FIELD_GET(IRDMA_CQ_MINERR, temp);
- if (info->error) {
- info->maj_err_code = (u16)FIELD_GET(IRDMA_CQ_MAJERR, temp);
- error = readl(cqp->dev->hw_regs[IRDMA_CQPERRCODES]);
- ibdev_dbg(to_ibdev(cqp->dev),
- "CQP: CQPERRCODES error_code[x%08X]\n", error);
- }
- wqe_idx = (u32)FIELD_GET(IRDMA_CQ_WQEIDX, temp);
- info->scratch = cqp->scratch_array[wqe_idx];
- get_64bit_val(cqe, 16, &temp1);
- info->op_ret_val = (u32)FIELD_GET(IRDMA_CCQ_OPRETVAL, temp1);
- get_64bit_val(cqp->sq_base[wqe_idx].elem, 24, &temp1);
- info->op_code = (u8)FIELD_GET(IRDMA_CQPSQ_OPCODE, temp1);
- info->cqp = cqp;
- /* move the head for cq */
- IRDMA_RING_MOVE_HEAD(ccq->cq_uk.cq_ring, ret_code);
- if (!IRDMA_RING_CURRENT_HEAD(ccq->cq_uk.cq_ring))
- ccq->cq_uk.polarity ^= 1;
- /* update cq tail in cq shadow memory also */
- IRDMA_RING_MOVE_TAIL(ccq->cq_uk.cq_ring);
- set_64bit_val(ccq->cq_uk.shadow_area, 0,
- IRDMA_RING_CURRENT_HEAD(ccq->cq_uk.cq_ring));
- dma_wmb(); /* make sure shadow area is updated before moving tail */
- IRDMA_RING_MOVE_TAIL(cqp->sq_ring);
- atomic64_inc(&cqp->completed_ops);
- return ret_code;
- }
- /**
- * irdma_sc_poll_for_cqp_op_done - Waits for last write to complete in CQP SQ
- * @cqp: struct for cqp hw
- * @op_code: cqp opcode for completion
- * @compl_info: completion q entry to return
- */
- int irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 op_code,
- struct irdma_ccq_cqe_info *compl_info)
- {
- struct irdma_ccq_cqe_info info = {};
- struct irdma_sc_cq *ccq;
- int ret_code = 0;
- u32 cnt = 0;
- ccq = cqp->dev->ccq;
- while (1) {
- if (cnt++ > 100 * cqp->dev->hw_attrs.max_done_count)
- return -ETIMEDOUT;
- if (irdma_sc_ccq_get_cqe_info(ccq, &info)) {
- udelay(cqp->dev->hw_attrs.max_sleep_count);
- continue;
- }
- if (info.error && info.op_code != IRDMA_CQP_OP_QUERY_STAG) {
- ret_code = -EIO;
- break;
- }
- /* make sure op code matches*/
- if (op_code == info.op_code)
- break;
- ibdev_dbg(to_ibdev(cqp->dev),
- "WQE: opcode mismatch for my op code 0x%x, returned opcode %x\n",
- op_code, info.op_code);
- }
- if (compl_info)
- memcpy(compl_info, &info, sizeof(*compl_info));
- return ret_code;
- }
- /**
- * irdma_sc_manage_hmc_pm_func_table - manage of function table
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @info: info for the manage function table operation
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_manage_hmc_pm_func_table(struct irdma_sc_cqp *cqp,
- struct irdma_hmc_fcn_info *info,
- u64 scratch, bool post_sq)
- {
- __le64 *wqe;
- u64 hdr;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 0, 0);
- set_64bit_val(wqe, 8, 0);
- set_64bit_val(wqe, 16, 0);
- set_64bit_val(wqe, 32, 0);
- set_64bit_val(wqe, 40, 0);
- set_64bit_val(wqe, 48, 0);
- set_64bit_val(wqe, 56, 0);
- hdr = FIELD_PREP(IRDMA_CQPSQ_MHMC_VFIDX, info->vf_id) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_MANAGE_HMC_PM_FUNC_TABLE) |
- FIELD_PREP(IRDMA_CQPSQ_MHMC_FREEPMFN, info->free_fcn) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: MANAGE_HMC_PM_FUNC_TABLE WQE",
- DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_commit_fpm_val_done - wait for cqp eqe completion
- * for fpm commit
- * @cqp: struct for cqp hw
- */
- static int irdma_sc_commit_fpm_val_done(struct irdma_sc_cqp *cqp)
- {
- return irdma_sc_poll_for_cqp_op_done(cqp, IRDMA_CQP_OP_COMMIT_FPM_VAL,
- NULL);
- }
- /**
- * irdma_sc_commit_fpm_val - cqp wqe for commit fpm values
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @hmc_fn_id: hmc function id
- * @commit_fpm_mem: Memory for fpm values
- * @post_sq: flag for cqp db to ring
- * @wait_type: poll ccq or cqp registers for cqp completion
- */
- static int irdma_sc_commit_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch,
- u8 hmc_fn_id,
- struct irdma_dma_mem *commit_fpm_mem,
- bool post_sq, u8 wait_type)
- {
- __le64 *wqe;
- u64 hdr;
- u32 tail, val, error;
- int ret_code = 0;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, hmc_fn_id);
- set_64bit_val(wqe, 32, commit_fpm_mem->pa);
- hdr = FIELD_PREP(IRDMA_CQPSQ_BUFSIZE, IRDMA_COMMIT_FPM_BUF_SIZE) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_COMMIT_FPM_VAL) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: COMMIT_FPM_VAL WQE", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_get_cqp_reg_info(cqp, &val, &tail, &error);
- if (post_sq) {
- irdma_sc_cqp_post_sq(cqp);
- if (wait_type == IRDMA_CQP_WAIT_POLL_REGS)
- ret_code = irdma_cqp_poll_registers(cqp, tail,
- cqp->dev->hw_attrs.max_done_count);
- else if (wait_type == IRDMA_CQP_WAIT_POLL_CQ)
- ret_code = irdma_sc_commit_fpm_val_done(cqp);
- }
- return ret_code;
- }
- /**
- * irdma_sc_query_fpm_val_done - poll for cqp wqe completion for
- * query fpm
- * @cqp: struct for cqp hw
- */
- static int irdma_sc_query_fpm_val_done(struct irdma_sc_cqp *cqp)
- {
- return irdma_sc_poll_for_cqp_op_done(cqp, IRDMA_CQP_OP_QUERY_FPM_VAL,
- NULL);
- }
- /**
- * irdma_sc_query_fpm_val - cqp wqe query fpm values
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @hmc_fn_id: hmc function id
- * @query_fpm_mem: memory for return fpm values
- * @post_sq: flag for cqp db to ring
- * @wait_type: poll ccq or cqp registers for cqp completion
- */
- static int irdma_sc_query_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch,
- u8 hmc_fn_id,
- struct irdma_dma_mem *query_fpm_mem,
- bool post_sq, u8 wait_type)
- {
- __le64 *wqe;
- u64 hdr;
- u32 tail, val, error;
- int ret_code = 0;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, hmc_fn_id);
- set_64bit_val(wqe, 32, query_fpm_mem->pa);
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_QUERY_FPM_VAL) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: QUERY_FPM WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_get_cqp_reg_info(cqp, &val, &tail, &error);
- if (post_sq) {
- irdma_sc_cqp_post_sq(cqp);
- if (wait_type == IRDMA_CQP_WAIT_POLL_REGS)
- ret_code = irdma_cqp_poll_registers(cqp, tail,
- cqp->dev->hw_attrs.max_done_count);
- else if (wait_type == IRDMA_CQP_WAIT_POLL_CQ)
- ret_code = irdma_sc_query_fpm_val_done(cqp);
- }
- return ret_code;
- }
- /**
- * irdma_sc_ceq_init - initialize ceq
- * @ceq: ceq sc structure
- * @info: ceq initialization info
- */
- int irdma_sc_ceq_init(struct irdma_sc_ceq *ceq,
- struct irdma_ceq_init_info *info)
- {
- u32 pble_obj_cnt;
- if (info->elem_cnt < info->dev->hw_attrs.min_hw_ceq_size ||
- info->elem_cnt > info->dev->hw_attrs.max_hw_ceq_size)
- return -EINVAL;
- if (info->ceq_id >= info->dev->hmc_fpm_misc.max_ceqs)
- return -EINVAL;
- pble_obj_cnt = info->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->virtual_map && info->first_pm_pbl_idx >= pble_obj_cnt)
- return -EINVAL;
- ceq->size = sizeof(*ceq);
- ceq->ceqe_base = (struct irdma_ceqe *)info->ceqe_base;
- ceq->ceq_id = info->ceq_id;
- ceq->dev = info->dev;
- ceq->elem_cnt = info->elem_cnt;
- ceq->ceq_elem_pa = info->ceqe_pa;
- ceq->virtual_map = info->virtual_map;
- ceq->itr_no_expire = info->itr_no_expire;
- ceq->reg_cq = info->reg_cq;
- ceq->reg_cq_size = 0;
- spin_lock_init(&ceq->req_cq_lock);
- ceq->pbl_chunk_size = (ceq->virtual_map ? info->pbl_chunk_size : 0);
- ceq->first_pm_pbl_idx = (ceq->virtual_map ? info->first_pm_pbl_idx : 0);
- ceq->pbl_list = (ceq->virtual_map ? info->pbl_list : NULL);
- ceq->tph_en = info->tph_en;
- ceq->tph_val = info->tph_val;
- ceq->vsi = info->vsi;
- ceq->polarity = 1;
- IRDMA_RING_INIT(ceq->ceq_ring, ceq->elem_cnt);
- ceq->dev->ceq[info->ceq_id] = ceq;
- return 0;
- }
- /**
- * irdma_sc_ceq_create - create ceq wqe
- * @ceq: ceq sc structure
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_ceq_create(struct irdma_sc_ceq *ceq, u64 scratch,
- bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- cqp = ceq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, ceq->elem_cnt);
- set_64bit_val(wqe, 32,
- (ceq->virtual_map ? 0 : ceq->ceq_elem_pa));
- set_64bit_val(wqe, 48,
- (ceq->virtual_map ? ceq->first_pm_pbl_idx : 0));
- set_64bit_val(wqe, 56,
- FIELD_PREP(IRDMA_CQPSQ_TPHVAL, ceq->tph_val) |
- FIELD_PREP(IRDMA_CQPSQ_VSIIDX, ceq->vsi->vsi_idx));
- hdr = FIELD_PREP(IRDMA_CQPSQ_CEQ_CEQID, ceq->ceq_id) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_CREATE_CEQ) |
- FIELD_PREP(IRDMA_CQPSQ_CEQ_LPBLSIZE, ceq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_CEQ_VMAP, ceq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_CEQ_ITRNOEXPIRE, ceq->itr_no_expire) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, ceq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CEQ_CREATE WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_cceq_create_done - poll for control ceq wqe to complete
- * @ceq: ceq sc structure
- */
- static int irdma_sc_cceq_create_done(struct irdma_sc_ceq *ceq)
- {
- struct irdma_sc_cqp *cqp;
- cqp = ceq->dev->cqp;
- return irdma_sc_poll_for_cqp_op_done(cqp, IRDMA_CQP_OP_CREATE_CEQ,
- NULL);
- }
- /**
- * irdma_sc_cceq_destroy_done - poll for destroy cceq to complete
- * @ceq: ceq sc structure
- */
- int irdma_sc_cceq_destroy_done(struct irdma_sc_ceq *ceq)
- {
- struct irdma_sc_cqp *cqp;
- if (ceq->reg_cq)
- irdma_sc_remove_cq_ctx(ceq, ceq->dev->ccq);
- cqp = ceq->dev->cqp;
- cqp->process_cqp_sds = irdma_update_sds_noccq;
- return irdma_sc_poll_for_cqp_op_done(cqp, IRDMA_CQP_OP_DESTROY_CEQ,
- NULL);
- }
- /**
- * irdma_sc_cceq_create - create cceq
- * @ceq: ceq sc structure
- * @scratch: u64 saved to be used during cqp completion
- */
- int irdma_sc_cceq_create(struct irdma_sc_ceq *ceq, u64 scratch)
- {
- int ret_code;
- struct irdma_sc_dev *dev = ceq->dev;
- dev->ccq->vsi = ceq->vsi;
- if (ceq->reg_cq) {
- ret_code = irdma_sc_add_cq_ctx(ceq, ceq->dev->ccq);
- if (ret_code)
- return ret_code;
- }
- ret_code = irdma_sc_ceq_create(ceq, scratch, true);
- if (!ret_code)
- return irdma_sc_cceq_create_done(ceq);
- return ret_code;
- }
- /**
- * irdma_sc_ceq_destroy - destroy ceq
- * @ceq: ceq sc structure
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_ceq_destroy(struct irdma_sc_ceq *ceq, u64 scratch, bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- cqp = ceq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, ceq->elem_cnt);
- set_64bit_val(wqe, 48, ceq->first_pm_pbl_idx);
- hdr = ceq->ceq_id |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DESTROY_CEQ) |
- FIELD_PREP(IRDMA_CQPSQ_CEQ_LPBLSIZE, ceq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_CEQ_VMAP, ceq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, ceq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CEQ_DESTROY WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_process_ceq - process ceq
- * @dev: sc device struct
- * @ceq: ceq sc structure
- *
- * It is expected caller serializes this function with cleanup_ceqes()
- * because these functions manipulate the same ceq
- */
- void *irdma_sc_process_ceq(struct irdma_sc_dev *dev, struct irdma_sc_ceq *ceq)
- {
- u64 temp;
- __le64 *ceqe;
- struct irdma_sc_cq *cq = NULL;
- struct irdma_sc_cq *temp_cq;
- u8 polarity;
- u32 cq_idx;
- unsigned long flags;
- do {
- cq_idx = 0;
- ceqe = IRDMA_GET_CURRENT_CEQ_ELEM(ceq);
- get_64bit_val(ceqe, 0, &temp);
- polarity = (u8)FIELD_GET(IRDMA_CEQE_VALID, temp);
- if (polarity != ceq->polarity)
- return NULL;
- temp_cq = (struct irdma_sc_cq *)(unsigned long)(temp << 1);
- if (!temp_cq) {
- cq_idx = IRDMA_INVALID_CQ_IDX;
- IRDMA_RING_MOVE_TAIL(ceq->ceq_ring);
- if (!IRDMA_RING_CURRENT_TAIL(ceq->ceq_ring))
- ceq->polarity ^= 1;
- continue;
- }
- cq = temp_cq;
- if (ceq->reg_cq) {
- spin_lock_irqsave(&ceq->req_cq_lock, flags);
- cq_idx = irdma_sc_find_reg_cq(ceq, cq);
- spin_unlock_irqrestore(&ceq->req_cq_lock, flags);
- }
- IRDMA_RING_MOVE_TAIL(ceq->ceq_ring);
- if (!IRDMA_RING_CURRENT_TAIL(ceq->ceq_ring))
- ceq->polarity ^= 1;
- } while (cq_idx == IRDMA_INVALID_CQ_IDX);
- if (cq)
- irdma_sc_cq_ack(cq);
- return cq;
- }
- /**
- * irdma_sc_cleanup_ceqes - clear the valid ceqes ctx matching the cq
- * @cq: cq for which the ceqes need to be cleaned up
- * @ceq: ceq ptr
- *
- * The function is called after the cq is destroyed to cleanup
- * its pending ceqe entries. It is expected caller serializes this
- * function with process_ceq() in interrupt context.
- */
- void irdma_sc_cleanup_ceqes(struct irdma_sc_cq *cq, struct irdma_sc_ceq *ceq)
- {
- struct irdma_sc_cq *next_cq;
- u8 ceq_polarity = ceq->polarity;
- __le64 *ceqe;
- u8 polarity;
- u64 temp;
- int next;
- u32 i;
- next = IRDMA_RING_GET_NEXT_TAIL(ceq->ceq_ring, 0);
- for (i = 1; i <= IRDMA_RING_SIZE(*ceq); i++) {
- ceqe = IRDMA_GET_CEQ_ELEM_AT_POS(ceq, next);
- get_64bit_val(ceqe, 0, &temp);
- polarity = (u8)FIELD_GET(IRDMA_CEQE_VALID, temp);
- if (polarity != ceq_polarity)
- return;
- next_cq = (struct irdma_sc_cq *)(unsigned long)(temp << 1);
- if (cq == next_cq)
- set_64bit_val(ceqe, 0, temp & IRDMA_CEQE_VALID);
- next = IRDMA_RING_GET_NEXT_TAIL(ceq->ceq_ring, i);
- if (!next)
- ceq_polarity ^= 1;
- }
- }
- /**
- * irdma_sc_aeq_init - initialize aeq
- * @aeq: aeq structure ptr
- * @info: aeq initialization info
- */
- int irdma_sc_aeq_init(struct irdma_sc_aeq *aeq,
- struct irdma_aeq_init_info *info)
- {
- u32 pble_obj_cnt;
- if (info->elem_cnt < info->dev->hw_attrs.min_hw_aeq_size ||
- info->elem_cnt > info->dev->hw_attrs.max_hw_aeq_size)
- return -EINVAL;
- pble_obj_cnt = info->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->virtual_map && info->first_pm_pbl_idx >= pble_obj_cnt)
- return -EINVAL;
- aeq->size = sizeof(*aeq);
- aeq->polarity = 1;
- aeq->aeqe_base = (struct irdma_sc_aeqe *)info->aeqe_base;
- aeq->dev = info->dev;
- aeq->elem_cnt = info->elem_cnt;
- aeq->aeq_elem_pa = info->aeq_elem_pa;
- IRDMA_RING_INIT(aeq->aeq_ring, aeq->elem_cnt);
- aeq->virtual_map = info->virtual_map;
- aeq->pbl_list = (aeq->virtual_map ? info->pbl_list : NULL);
- aeq->pbl_chunk_size = (aeq->virtual_map ? info->pbl_chunk_size : 0);
- aeq->first_pm_pbl_idx = (aeq->virtual_map ? info->first_pm_pbl_idx : 0);
- aeq->msix_idx = info->msix_idx;
- info->dev->aeq = aeq;
- return 0;
- }
- /**
- * irdma_sc_aeq_create - create aeq
- * @aeq: aeq structure ptr
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_aeq_create(struct irdma_sc_aeq *aeq, u64 scratch,
- bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- u64 hdr;
- cqp = aeq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, aeq->elem_cnt);
- set_64bit_val(wqe, 32,
- (aeq->virtual_map ? 0 : aeq->aeq_elem_pa));
- set_64bit_val(wqe, 48,
- (aeq->virtual_map ? aeq->first_pm_pbl_idx : 0));
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_CREATE_AEQ) |
- FIELD_PREP(IRDMA_CQPSQ_AEQ_LPBLSIZE, aeq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_AEQ_VMAP, aeq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: AEQ_CREATE WQE", DUMP_PREFIX_OFFSET, 16, 8,
- wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_aeq_destroy - destroy aeq during close
- * @aeq: aeq structure ptr
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- static int irdma_sc_aeq_destroy(struct irdma_sc_aeq *aeq, u64 scratch,
- bool post_sq)
- {
- __le64 *wqe;
- struct irdma_sc_cqp *cqp;
- struct irdma_sc_dev *dev;
- u64 hdr;
- dev = aeq->dev;
- writel(0, dev->hw_regs[IRDMA_PFINT_AEQCTL]);
- cqp = dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16, aeq->elem_cnt);
- set_64bit_val(wqe, 48, aeq->first_pm_pbl_idx);
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DESTROY_AEQ) |
- FIELD_PREP(IRDMA_CQPSQ_AEQ_LPBLSIZE, aeq->pbl_chunk_size) |
- FIELD_PREP(IRDMA_CQPSQ_AEQ_VMAP, aeq->virtual_map) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: AEQ_DESTROY WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- if (post_sq)
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_sc_get_next_aeqe - get next aeq entry
- * @aeq: aeq structure ptr
- * @info: aeqe info to be returned
- */
- int irdma_sc_get_next_aeqe(struct irdma_sc_aeq *aeq,
- struct irdma_aeqe_info *info)
- {
- u64 temp, compl_ctx;
- __le64 *aeqe;
- u16 wqe_idx;
- u8 ae_src;
- u8 polarity;
- aeqe = IRDMA_GET_CURRENT_AEQ_ELEM(aeq);
- get_64bit_val(aeqe, 8, &temp);
- polarity = (u8)FIELD_GET(IRDMA_AEQE_VALID, temp);
- if (aeq->polarity != polarity)
- return -ENOENT;
- /* Ensure AEQE contents are read after valid bit is checked */
- dma_rmb();
- get_64bit_val(aeqe, 0, &compl_ctx);
- print_hex_dump_debug("WQE: AEQ_ENTRY WQE", DUMP_PREFIX_OFFSET, 16, 8,
- aeqe, 16, false);
- ae_src = (u8)FIELD_GET(IRDMA_AEQE_AESRC, temp);
- wqe_idx = (u16)FIELD_GET(IRDMA_AEQE_WQDESCIDX, temp);
- info->qp_cq_id = (u32)FIELD_GET(IRDMA_AEQE_QPCQID_LOW, temp) |
- ((u32)FIELD_GET(IRDMA_AEQE_QPCQID_HI, temp) << 18);
- info->ae_id = (u16)FIELD_GET(IRDMA_AEQE_AECODE, temp);
- info->tcp_state = (u8)FIELD_GET(IRDMA_AEQE_TCPSTATE, temp);
- info->iwarp_state = (u8)FIELD_GET(IRDMA_AEQE_IWSTATE, temp);
- info->q2_data_written = (u8)FIELD_GET(IRDMA_AEQE_Q2DATA, temp);
- info->aeqe_overflow = (bool)FIELD_GET(IRDMA_AEQE_OVERFLOW, temp);
- info->ae_src = ae_src;
- switch (info->ae_id) {
- case IRDMA_AE_PRIV_OPERATION_DENIED:
- case IRDMA_AE_AMP_INVALIDATE_TYPE1_MW:
- case IRDMA_AE_AMP_MWBIND_ZERO_BASED_TYPE1_MW:
- case IRDMA_AE_AMP_FASTREG_INVALID_PBL_HPS_CFG:
- case IRDMA_AE_AMP_FASTREG_PBLE_MISMATCH:
- case IRDMA_AE_UDA_XMIT_DGRAM_TOO_LONG:
- case IRDMA_AE_UDA_XMIT_BAD_PD:
- case IRDMA_AE_UDA_XMIT_DGRAM_TOO_SHORT:
- case IRDMA_AE_BAD_CLOSE:
- case IRDMA_AE_RDMA_READ_WHILE_ORD_ZERO:
- case IRDMA_AE_STAG_ZERO_INVALID:
- case IRDMA_AE_IB_RREQ_AND_Q1_FULL:
- case IRDMA_AE_IB_INVALID_REQUEST:
- case IRDMA_AE_WQE_UNEXPECTED_OPCODE:
- case IRDMA_AE_IB_REMOTE_ACCESS_ERROR:
- case IRDMA_AE_IB_REMOTE_OP_ERROR:
- case IRDMA_AE_DDP_UBE_INVALID_DDP_VERSION:
- case IRDMA_AE_DDP_UBE_INVALID_MO:
- case IRDMA_AE_DDP_UBE_INVALID_QN:
- case IRDMA_AE_DDP_NO_L_BIT:
- case IRDMA_AE_RDMAP_ROE_INVALID_RDMAP_VERSION:
- case IRDMA_AE_RDMAP_ROE_UNEXPECTED_OPCODE:
- case IRDMA_AE_ROE_INVALID_RDMA_READ_REQUEST:
- case IRDMA_AE_ROE_INVALID_RDMA_WRITE_OR_READ_RESP:
- case IRDMA_AE_ROCE_RSP_LENGTH_ERROR:
- case IRDMA_AE_INVALID_ARP_ENTRY:
- case IRDMA_AE_INVALID_TCP_OPTION_RCVD:
- case IRDMA_AE_STALE_ARP_ENTRY:
- case IRDMA_AE_INVALID_AH_ENTRY:
- case IRDMA_AE_LLP_RECEIVED_MPA_CRC_ERROR:
- case IRDMA_AE_LLP_SEGMENT_TOO_SMALL:
- case IRDMA_AE_LLP_TOO_MANY_RETRIES:
- case IRDMA_AE_LLP_DOUBT_REACHABILITY:
- case IRDMA_AE_LLP_CONNECTION_ESTABLISHED:
- case IRDMA_AE_RESET_SENT:
- case IRDMA_AE_TERMINATE_SENT:
- case IRDMA_AE_RESET_NOT_SENT:
- case IRDMA_AE_LCE_QP_CATASTROPHIC:
- case IRDMA_AE_QP_SUSPEND_COMPLETE:
- case IRDMA_AE_UDA_L4LEN_INVALID:
- info->qp = true;
- info->compl_ctx = compl_ctx;
- break;
- case IRDMA_AE_LCE_CQ_CATASTROPHIC:
- info->cq = true;
- info->compl_ctx = compl_ctx << 1;
- ae_src = IRDMA_AE_SOURCE_RSVD;
- break;
- case IRDMA_AE_ROCE_EMPTY_MCG:
- case IRDMA_AE_ROCE_BAD_MC_IP_ADDR:
- case IRDMA_AE_ROCE_BAD_MC_QPID:
- case IRDMA_AE_MCG_QP_PROTOCOL_MISMATCH:
- fallthrough;
- case IRDMA_AE_LLP_CONNECTION_RESET:
- case IRDMA_AE_LLP_SYN_RECEIVED:
- case IRDMA_AE_LLP_FIN_RECEIVED:
- case IRDMA_AE_LLP_CLOSE_COMPLETE:
- case IRDMA_AE_LLP_TERMINATE_RECEIVED:
- case IRDMA_AE_RDMAP_ROE_BAD_LLP_CLOSE:
- ae_src = IRDMA_AE_SOURCE_RSVD;
- info->qp = true;
- info->compl_ctx = compl_ctx;
- break;
- default:
- break;
- }
- switch (ae_src) {
- case IRDMA_AE_SOURCE_RQ:
- case IRDMA_AE_SOURCE_RQ_0011:
- info->qp = true;
- info->rq = true;
- info->wqe_idx = wqe_idx;
- info->compl_ctx = compl_ctx;
- break;
- case IRDMA_AE_SOURCE_CQ:
- case IRDMA_AE_SOURCE_CQ_0110:
- case IRDMA_AE_SOURCE_CQ_1010:
- case IRDMA_AE_SOURCE_CQ_1110:
- info->cq = true;
- info->compl_ctx = compl_ctx << 1;
- break;
- case IRDMA_AE_SOURCE_SQ:
- case IRDMA_AE_SOURCE_SQ_0111:
- info->qp = true;
- info->sq = true;
- info->wqe_idx = wqe_idx;
- info->compl_ctx = compl_ctx;
- break;
- case IRDMA_AE_SOURCE_IN_RR_WR:
- case IRDMA_AE_SOURCE_IN_RR_WR_1011:
- info->qp = true;
- info->compl_ctx = compl_ctx;
- info->in_rdrsp_wr = true;
- break;
- case IRDMA_AE_SOURCE_OUT_RR:
- case IRDMA_AE_SOURCE_OUT_RR_1111:
- info->qp = true;
- info->compl_ctx = compl_ctx;
- info->out_rdrsp = true;
- break;
- case IRDMA_AE_SOURCE_RSVD:
- default:
- break;
- }
- IRDMA_RING_MOVE_TAIL(aeq->aeq_ring);
- if (!IRDMA_RING_CURRENT_TAIL(aeq->aeq_ring))
- aeq->polarity ^= 1;
- return 0;
- }
- /**
- * irdma_sc_repost_aeq_entries - repost completed aeq entries
- * @dev: sc device struct
- * @count: allocate count
- */
- void irdma_sc_repost_aeq_entries(struct irdma_sc_dev *dev, u32 count)
- {
- writel(count, dev->hw_regs[IRDMA_AEQALLOC]);
- }
- /**
- * irdma_sc_ccq_init - initialize control cq
- * @cq: sc's cq ctruct
- * @info: info for control cq initialization
- */
- int irdma_sc_ccq_init(struct irdma_sc_cq *cq, struct irdma_ccq_init_info *info)
- {
- u32 pble_obj_cnt;
- if (info->num_elem < info->dev->hw_attrs.uk_attrs.min_hw_cq_size ||
- info->num_elem > info->dev->hw_attrs.uk_attrs.max_hw_cq_size)
- return -EINVAL;
- if (info->ceq_id >= info->dev->hmc_fpm_misc.max_ceqs)
- return -EINVAL;
- pble_obj_cnt = info->dev->hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt;
- if (info->virtual_map && info->first_pm_pbl_idx >= pble_obj_cnt)
- return -EINVAL;
- cq->cq_pa = info->cq_pa;
- cq->cq_uk.cq_base = info->cq_base;
- cq->shadow_area_pa = info->shadow_area_pa;
- cq->cq_uk.shadow_area = info->shadow_area;
- cq->shadow_read_threshold = info->shadow_read_threshold;
- cq->dev = info->dev;
- cq->ceq_id = info->ceq_id;
- cq->cq_uk.cq_size = info->num_elem;
- cq->cq_type = IRDMA_CQ_TYPE_CQP;
- cq->ceqe_mask = info->ceqe_mask;
- IRDMA_RING_INIT(cq->cq_uk.cq_ring, info->num_elem);
- cq->cq_uk.cq_id = 0; /* control cq is id 0 always */
- cq->ceq_id_valid = info->ceq_id_valid;
- cq->tph_en = info->tph_en;
- cq->tph_val = info->tph_val;
- cq->cq_uk.avoid_mem_cflct = info->avoid_mem_cflct;
- cq->pbl_list = info->pbl_list;
- cq->virtual_map = info->virtual_map;
- cq->pbl_chunk_size = info->pbl_chunk_size;
- cq->first_pm_pbl_idx = info->first_pm_pbl_idx;
- cq->cq_uk.polarity = true;
- cq->vsi = info->vsi;
- cq->cq_uk.cq_ack_db = cq->dev->cq_ack_db;
- /* Only applicable to CQs other than CCQ so initialize to zero */
- cq->cq_uk.cqe_alloc_db = NULL;
- info->dev->ccq = cq;
- return 0;
- }
- /**
- * irdma_sc_ccq_create_done - poll cqp for ccq create
- * @ccq: ccq sc struct
- */
- static inline int irdma_sc_ccq_create_done(struct irdma_sc_cq *ccq)
- {
- struct irdma_sc_cqp *cqp;
- cqp = ccq->dev->cqp;
- return irdma_sc_poll_for_cqp_op_done(cqp, IRDMA_CQP_OP_CREATE_CQ, NULL);
- }
- /**
- * irdma_sc_ccq_create - create control cq
- * @ccq: ccq sc struct
- * @scratch: u64 saved to be used during cqp completion
- * @check_overflow: overlow flag for ccq
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_ccq_create(struct irdma_sc_cq *ccq, u64 scratch,
- bool check_overflow, bool post_sq)
- {
- int ret_code;
- ret_code = irdma_sc_cq_create(ccq, scratch, check_overflow, post_sq);
- if (ret_code)
- return ret_code;
- if (post_sq) {
- ret_code = irdma_sc_ccq_create_done(ccq);
- if (ret_code)
- return ret_code;
- }
- ccq->dev->cqp->process_cqp_sds = irdma_cqp_sds_cmd;
- return 0;
- }
- /**
- * irdma_sc_ccq_destroy - destroy ccq during close
- * @ccq: ccq sc struct
- * @scratch: u64 saved to be used during cqp completion
- * @post_sq: flag for cqp db to ring
- */
- int irdma_sc_ccq_destroy(struct irdma_sc_cq *ccq, u64 scratch, bool post_sq)
- {
- struct irdma_sc_cqp *cqp;
- __le64 *wqe;
- u64 hdr;
- int ret_code = 0;
- u32 tail, val, error;
- cqp = ccq->dev->cqp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 0, ccq->cq_uk.cq_size);
- set_64bit_val(wqe, 8, (uintptr_t)ccq >> 1);
- set_64bit_val(wqe, 40, ccq->shadow_area_pa);
- hdr = ccq->cq_uk.cq_id |
- FLD_LS_64(ccq->dev, (ccq->ceq_id_valid ? ccq->ceq_id : 0),
- IRDMA_CQPSQ_CQ_CEQID) |
- FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_DESTROY_CQ) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_ENCEQEMASK, ccq->ceqe_mask) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_CEQIDVALID, ccq->ceq_id_valid) |
- FIELD_PREP(IRDMA_CQPSQ_TPHEN, ccq->tph_en) |
- FIELD_PREP(IRDMA_CQPSQ_CQ_AVOIDMEMCNFLCT, ccq->cq_uk.avoid_mem_cflct) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: CCQ_DESTROY WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_get_cqp_reg_info(cqp, &val, &tail, &error);
- if (post_sq) {
- irdma_sc_cqp_post_sq(cqp);
- ret_code = irdma_cqp_poll_registers(cqp, tail,
- cqp->dev->hw_attrs.max_done_count);
- }
- cqp->process_cqp_sds = irdma_update_sds_noccq;
- return ret_code;
- }
- /**
- * irdma_sc_init_iw_hmc() - queries fpm values using cqp and populates hmc_info
- * @dev : ptr to irdma_dev struct
- * @hmc_fn_id: hmc function id
- */
- int irdma_sc_init_iw_hmc(struct irdma_sc_dev *dev, u8 hmc_fn_id)
- {
- struct irdma_hmc_info *hmc_info;
- struct irdma_hmc_fpm_misc *hmc_fpm_misc;
- struct irdma_dma_mem query_fpm_mem;
- int ret_code = 0;
- u8 wait_type;
- hmc_info = dev->hmc_info;
- hmc_fpm_misc = &dev->hmc_fpm_misc;
- query_fpm_mem.pa = dev->fpm_query_buf_pa;
- query_fpm_mem.va = dev->fpm_query_buf;
- hmc_info->hmc_fn_id = hmc_fn_id;
- wait_type = (u8)IRDMA_CQP_WAIT_POLL_REGS;
- ret_code = irdma_sc_query_fpm_val(dev->cqp, 0, hmc_info->hmc_fn_id,
- &query_fpm_mem, true, wait_type);
- if (ret_code)
- return ret_code;
- /* parse the fpm_query_buf and fill hmc obj info */
- ret_code = irdma_sc_parse_fpm_query_buf(dev, query_fpm_mem.va, hmc_info,
- hmc_fpm_misc);
- print_hex_dump_debug("HMC: QUERY FPM BUFFER", DUMP_PREFIX_OFFSET, 16,
- 8, query_fpm_mem.va, IRDMA_QUERY_FPM_BUF_SIZE,
- false);
- return ret_code;
- }
- /**
- * irdma_sc_cfg_iw_fpm() - commits hmc obj cnt values using cqp
- * command and populates fpm base address in hmc_info
- * @dev : ptr to irdma_dev struct
- * @hmc_fn_id: hmc function id
- */
- static int irdma_sc_cfg_iw_fpm(struct irdma_sc_dev *dev, u8 hmc_fn_id)
- {
- struct irdma_hmc_info *hmc_info;
- struct irdma_hmc_obj_info *obj_info;
- __le64 *buf;
- struct irdma_dma_mem commit_fpm_mem;
- int ret_code = 0;
- u8 wait_type;
- hmc_info = dev->hmc_info;
- obj_info = hmc_info->hmc_obj;
- buf = dev->fpm_commit_buf;
- set_64bit_val(buf, 0, (u64)obj_info[IRDMA_HMC_IW_QP].cnt);
- set_64bit_val(buf, 8, (u64)obj_info[IRDMA_HMC_IW_CQ].cnt);
- set_64bit_val(buf, 16, (u64)0); /* RSRVD */
- set_64bit_val(buf, 24, (u64)obj_info[IRDMA_HMC_IW_HTE].cnt);
- set_64bit_val(buf, 32, (u64)obj_info[IRDMA_HMC_IW_ARP].cnt);
- set_64bit_val(buf, 40, (u64)0); /* RSVD */
- set_64bit_val(buf, 48, (u64)obj_info[IRDMA_HMC_IW_MR].cnt);
- set_64bit_val(buf, 56, (u64)obj_info[IRDMA_HMC_IW_XF].cnt);
- set_64bit_val(buf, 64, (u64)obj_info[IRDMA_HMC_IW_XFFL].cnt);
- set_64bit_val(buf, 72, (u64)obj_info[IRDMA_HMC_IW_Q1].cnt);
- set_64bit_val(buf, 80, (u64)obj_info[IRDMA_HMC_IW_Q1FL].cnt);
- set_64bit_val(buf, 88,
- (u64)obj_info[IRDMA_HMC_IW_TIMER].cnt);
- set_64bit_val(buf, 96,
- (u64)obj_info[IRDMA_HMC_IW_FSIMC].cnt);
- set_64bit_val(buf, 104,
- (u64)obj_info[IRDMA_HMC_IW_FSIAV].cnt);
- set_64bit_val(buf, 112,
- (u64)obj_info[IRDMA_HMC_IW_PBLE].cnt);
- set_64bit_val(buf, 120, (u64)0); /* RSVD */
- set_64bit_val(buf, 128, (u64)obj_info[IRDMA_HMC_IW_RRF].cnt);
- set_64bit_val(buf, 136,
- (u64)obj_info[IRDMA_HMC_IW_RRFFL].cnt);
- set_64bit_val(buf, 144, (u64)obj_info[IRDMA_HMC_IW_HDR].cnt);
- set_64bit_val(buf, 152, (u64)obj_info[IRDMA_HMC_IW_MD].cnt);
- set_64bit_val(buf, 160,
- (u64)obj_info[IRDMA_HMC_IW_OOISC].cnt);
- set_64bit_val(buf, 168,
- (u64)obj_info[IRDMA_HMC_IW_OOISCFFL].cnt);
- commit_fpm_mem.pa = dev->fpm_commit_buf_pa;
- commit_fpm_mem.va = dev->fpm_commit_buf;
- wait_type = (u8)IRDMA_CQP_WAIT_POLL_REGS;
- print_hex_dump_debug("HMC: COMMIT FPM BUFFER", DUMP_PREFIX_OFFSET, 16,
- 8, commit_fpm_mem.va, IRDMA_COMMIT_FPM_BUF_SIZE,
- false);
- ret_code = irdma_sc_commit_fpm_val(dev->cqp, 0, hmc_info->hmc_fn_id,
- &commit_fpm_mem, true, wait_type);
- if (!ret_code)
- irdma_sc_parse_fpm_commit_buf(dev, dev->fpm_commit_buf,
- hmc_info->hmc_obj,
- &hmc_info->sd_table.sd_cnt);
- print_hex_dump_debug("HMC: COMMIT FPM BUFFER", DUMP_PREFIX_OFFSET, 16,
- 8, commit_fpm_mem.va, IRDMA_COMMIT_FPM_BUF_SIZE,
- false);
- return ret_code;
- }
- /**
- * cqp_sds_wqe_fill - fill cqp wqe doe sd
- * @cqp: struct for cqp hw
- * @info: sd info for wqe
- * @scratch: u64 saved to be used during cqp completion
- */
- static int cqp_sds_wqe_fill(struct irdma_sc_cqp *cqp,
- struct irdma_update_sds_info *info, u64 scratch)
- {
- u64 data;
- u64 hdr;
- __le64 *wqe;
- int mem_entries, wqe_entries;
- struct irdma_dma_mem *sdbuf = &cqp->sdbuf;
- u64 offset = 0;
- u32 wqe_idx;
- wqe = irdma_sc_cqp_get_next_send_wqe_idx(cqp, scratch, &wqe_idx);
- if (!wqe)
- return -ENOMEM;
- wqe_entries = (info->cnt > 3) ? 3 : info->cnt;
- mem_entries = info->cnt - wqe_entries;
- if (mem_entries) {
- offset = wqe_idx * IRDMA_UPDATE_SD_BUFF_SIZE;
- memcpy(((char *)sdbuf->va + offset), &info->entry[3], mem_entries << 4);
- data = (u64)sdbuf->pa + offset;
- } else {
- data = 0;
- }
- data |= FIELD_PREP(IRDMA_CQPSQ_UPESD_HMCFNID, info->hmc_fn_id);
- set_64bit_val(wqe, 16, data);
- switch (wqe_entries) {
- case 3:
- set_64bit_val(wqe, 48,
- (FIELD_PREP(IRDMA_CQPSQ_UPESD_SDCMD, info->entry[2].cmd) |
- FIELD_PREP(IRDMA_CQPSQ_UPESD_ENTRY_VALID, 1)));
- set_64bit_val(wqe, 56, info->entry[2].data);
- fallthrough;
- case 2:
- set_64bit_val(wqe, 32,
- (FIELD_PREP(IRDMA_CQPSQ_UPESD_SDCMD, info->entry[1].cmd) |
- FIELD_PREP(IRDMA_CQPSQ_UPESD_ENTRY_VALID, 1)));
- set_64bit_val(wqe, 40, info->entry[1].data);
- fallthrough;
- case 1:
- set_64bit_val(wqe, 0,
- FIELD_PREP(IRDMA_CQPSQ_UPESD_SDCMD, info->entry[0].cmd));
- set_64bit_val(wqe, 8, info->entry[0].data);
- break;
- default:
- break;
- }
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE, IRDMA_CQP_OP_UPDATE_PE_SDS) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_UPESD_ENTRY_COUNT, mem_entries);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- if (mem_entries)
- print_hex_dump_debug("WQE: UPDATE_PE_SDS WQE Buffer",
- DUMP_PREFIX_OFFSET, 16, 8,
- (char *)sdbuf->va + offset,
- mem_entries << 4, false);
- print_hex_dump_debug("WQE: UPDATE_PE_SDS WQE", DUMP_PREFIX_OFFSET, 16,
- 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- return 0;
- }
- /**
- * irdma_update_pe_sds - cqp wqe for sd
- * @dev: ptr to irdma_dev struct
- * @info: sd info for sd's
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_update_pe_sds(struct irdma_sc_dev *dev,
- struct irdma_update_sds_info *info, u64 scratch)
- {
- struct irdma_sc_cqp *cqp = dev->cqp;
- int ret_code;
- ret_code = cqp_sds_wqe_fill(cqp, info, scratch);
- if (!ret_code)
- irdma_sc_cqp_post_sq(cqp);
- return ret_code;
- }
- /**
- * irdma_update_sds_noccq - update sd before ccq created
- * @dev: sc device struct
- * @info: sd info for sd's
- */
- int irdma_update_sds_noccq(struct irdma_sc_dev *dev,
- struct irdma_update_sds_info *info)
- {
- u32 error, val, tail;
- struct irdma_sc_cqp *cqp = dev->cqp;
- int ret_code;
- ret_code = cqp_sds_wqe_fill(cqp, info, 0);
- if (ret_code)
- return ret_code;
- irdma_get_cqp_reg_info(cqp, &val, &tail, &error);
- irdma_sc_cqp_post_sq(cqp);
- return irdma_cqp_poll_registers(cqp, tail,
- cqp->dev->hw_attrs.max_done_count);
- }
- /**
- * irdma_sc_static_hmc_pages_allocated - cqp wqe to allocate hmc pages
- * @cqp: struct for cqp hw
- * @scratch: u64 saved to be used during cqp completion
- * @hmc_fn_id: hmc function id
- * @post_sq: flag for cqp db to ring
- * @poll_registers: flag to poll register for cqp completion
- */
- int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
- u8 hmc_fn_id, bool post_sq,
- bool poll_registers)
- {
- u64 hdr;
- __le64 *wqe;
- u32 tail, val, error;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- set_64bit_val(wqe, 16,
- FIELD_PREP(IRDMA_SHMC_PAGE_ALLOCATED_HMC_FN_ID, hmc_fn_id));
- hdr = FIELD_PREP(IRDMA_CQPSQ_OPCODE,
- IRDMA_CQP_OP_SHMC_PAGES_ALLOCATED) |
- FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, hdr);
- print_hex_dump_debug("WQE: SHMC_PAGES_ALLOCATED WQE",
- DUMP_PREFIX_OFFSET, 16, 8, wqe,
- IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_get_cqp_reg_info(cqp, &val, &tail, &error);
- if (post_sq) {
- irdma_sc_cqp_post_sq(cqp);
- if (poll_registers)
- /* check for cqp sq tail update */
- return irdma_cqp_poll_registers(cqp, tail,
- cqp->dev->hw_attrs.max_done_count);
- else
- return irdma_sc_poll_for_cqp_op_done(cqp,
- IRDMA_CQP_OP_SHMC_PAGES_ALLOCATED,
- NULL);
- }
- return 0;
- }
- /**
- * irdma_cqp_ring_full - check if cqp ring is full
- * @cqp: struct for cqp hw
- */
- static bool irdma_cqp_ring_full(struct irdma_sc_cqp *cqp)
- {
- return IRDMA_RING_FULL_ERR(cqp->sq_ring);
- }
- /**
- * irdma_est_sd - returns approximate number of SDs for HMC
- * @dev: sc device struct
- * @hmc_info: hmc structure, size and count for HMC objects
- */
- static u32 irdma_est_sd(struct irdma_sc_dev *dev,
- struct irdma_hmc_info *hmc_info)
- {
- int i;
- u64 size = 0;
- u64 sd;
- for (i = IRDMA_HMC_IW_QP; i < IRDMA_HMC_IW_MAX; i++)
- if (i != IRDMA_HMC_IW_PBLE)
- size += round_up(hmc_info->hmc_obj[i].cnt *
- hmc_info->hmc_obj[i].size, 512);
- size += round_up(hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt *
- hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].size, 512);
- if (size & 0x1FFFFF)
- sd = (size >> 21) + 1; /* add 1 for remainder */
- else
- sd = size >> 21;
- if (sd > 0xFFFFFFFF) {
- ibdev_dbg(to_ibdev(dev), "HMC: sd overflow[%lld]\n", sd);
- sd = 0xFFFFFFFF - 1;
- }
- return (u32)sd;
- }
- /**
- * irdma_sc_query_rdma_features_done - poll cqp for query features done
- * @cqp: struct for cqp hw
- */
- static int irdma_sc_query_rdma_features_done(struct irdma_sc_cqp *cqp)
- {
- return irdma_sc_poll_for_cqp_op_done(cqp,
- IRDMA_CQP_OP_QUERY_RDMA_FEATURES,
- NULL);
- }
- /**
- * irdma_sc_query_rdma_features - query RDMA features and FW ver
- * @cqp: struct for cqp hw
- * @buf: buffer to hold query info
- * @scratch: u64 saved to be used during cqp completion
- */
- static int irdma_sc_query_rdma_features(struct irdma_sc_cqp *cqp,
- struct irdma_dma_mem *buf, u64 scratch)
- {
- __le64 *wqe;
- u64 temp;
- wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch);
- if (!wqe)
- return -ENOMEM;
- temp = buf->pa;
- set_64bit_val(wqe, 32, temp);
- temp = FIELD_PREP(IRDMA_CQPSQ_QUERY_RDMA_FEATURES_WQEVALID,
- cqp->polarity) |
- FIELD_PREP(IRDMA_CQPSQ_QUERY_RDMA_FEATURES_BUF_LEN, buf->size) |
- FIELD_PREP(IRDMA_CQPSQ_UP_OP, IRDMA_CQP_OP_QUERY_RDMA_FEATURES);
- dma_wmb(); /* make sure WQE is written before valid bit is set */
- set_64bit_val(wqe, 24, temp);
- print_hex_dump_debug("WQE: QUERY RDMA FEATURES", DUMP_PREFIX_OFFSET,
- 16, 8, wqe, IRDMA_CQP_WQE_SIZE * 8, false);
- irdma_sc_cqp_post_sq(cqp);
- return 0;
- }
- /**
- * irdma_get_rdma_features - get RDMA features
- * @dev: sc device struct
- */
- int irdma_get_rdma_features(struct irdma_sc_dev *dev)
- {
- int ret_code;
- struct irdma_dma_mem feat_buf;
- u64 temp;
- u16 byte_idx, feat_type, feat_cnt, feat_idx;
- feat_buf.size = ALIGN(IRDMA_FEATURE_BUF_SIZE,
- IRDMA_FEATURE_BUF_ALIGNMENT);
- feat_buf.va = dma_alloc_coherent(dev->hw->device, feat_buf.size,
- &feat_buf.pa, GFP_KERNEL);
- if (!feat_buf.va)
- return -ENOMEM;
- ret_code = irdma_sc_query_rdma_features(dev->cqp, &feat_buf, 0);
- if (!ret_code)
- ret_code = irdma_sc_query_rdma_features_done(dev->cqp);
- if (ret_code)
- goto exit;
- get_64bit_val(feat_buf.va, 0, &temp);
- feat_cnt = (u16)FIELD_GET(IRDMA_FEATURE_CNT, temp);
- if (feat_cnt < 2) {
- ret_code = -EINVAL;
- goto exit;
- } else if (feat_cnt > IRDMA_MAX_FEATURES) {
- ibdev_dbg(to_ibdev(dev),
- "DEV: feature buf size insufficient, retrying with larger buffer\n");
- dma_free_coherent(dev->hw->device, feat_buf.size, feat_buf.va,
- feat_buf.pa);
- feat_buf.va = NULL;
- feat_buf.size = ALIGN(8 * feat_cnt,
- IRDMA_FEATURE_BUF_ALIGNMENT);
- feat_buf.va = dma_alloc_coherent(dev->hw->device,
- feat_buf.size, &feat_buf.pa,
- GFP_KERNEL);
- if (!feat_buf.va)
- return -ENOMEM;
- ret_code = irdma_sc_query_rdma_features(dev->cqp, &feat_buf, 0);
- if (!ret_code)
- ret_code = irdma_sc_query_rdma_features_done(dev->cqp);
- if (ret_code)
- goto exit;
- get_64bit_val(feat_buf.va, 0, &temp);
- feat_cnt = (u16)FIELD_GET(IRDMA_FEATURE_CNT, temp);
- if (feat_cnt < 2) {
- ret_code = -EINVAL;
- goto exit;
- }
- }
- print_hex_dump_debug("WQE: QUERY RDMA FEATURES", DUMP_PREFIX_OFFSET,
- 16, 8, feat_buf.va, feat_cnt * 8, false);
- for (byte_idx = 0, feat_idx = 0; feat_idx < min(feat_cnt, (u16)IRDMA_MAX_FEATURES);
- feat_idx++, byte_idx += 8) {
- get_64bit_val(feat_buf.va, byte_idx, &temp);
- feat_type = FIELD_GET(IRDMA_FEATURE_TYPE, temp);
- if (feat_type >= IRDMA_MAX_FEATURES) {
- ibdev_dbg(to_ibdev(dev),
- "DEV: found unrecognized feature type %d\n",
- feat_type);
- continue;
- }
- dev->feature_info[feat_type] = temp;
- }
- exit:
- dma_free_coherent(dev->hw->device, feat_buf.size, feat_buf.va,
- feat_buf.pa);
- feat_buf.va = NULL;
- return ret_code;
- }
- static u32 irdma_q1_cnt(struct irdma_sc_dev *dev,
- struct irdma_hmc_info *hmc_info, u32 qpwanted)
- {
- u32 q1_cnt;
- if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) {
- q1_cnt = roundup_pow_of_two(dev->hw_attrs.max_hw_ird * 2 * qpwanted);
- } else {
- if (dev->cqp->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
- q1_cnt = roundup_pow_of_two(dev->hw_attrs.max_hw_ird * 2 * qpwanted + 512);
- else
- q1_cnt = dev->hw_attrs.max_hw_ird * 2 * qpwanted;
- }
- return q1_cnt;
- }
- static void cfg_fpm_value_gen_1(struct irdma_sc_dev *dev,
- struct irdma_hmc_info *hmc_info, u32 qpwanted)
- {
- hmc_info->hmc_obj[IRDMA_HMC_IW_XF].cnt = roundup_pow_of_two(qpwanted * dev->hw_attrs.max_hw_wqes);
- }
- static void cfg_fpm_value_gen_2(struct irdma_sc_dev *dev,
- struct irdma_hmc_info *hmc_info, u32 qpwanted)
- {
- struct irdma_hmc_fpm_misc *hmc_fpm_misc = &dev->hmc_fpm_misc;
- hmc_info->hmc_obj[IRDMA_HMC_IW_XF].cnt =
- 4 * hmc_fpm_misc->xf_block_size * qpwanted;
- hmc_info->hmc_obj[IRDMA_HMC_IW_HDR].cnt = qpwanted;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_RRF].max_cnt)
- hmc_info->hmc_obj[IRDMA_HMC_IW_RRF].cnt = 32 * qpwanted;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_RRFFL].max_cnt)
- hmc_info->hmc_obj[IRDMA_HMC_IW_RRFFL].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_RRF].cnt /
- hmc_fpm_misc->rrf_block_size;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_OOISC].max_cnt)
- hmc_info->hmc_obj[IRDMA_HMC_IW_OOISC].cnt = 32 * qpwanted;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_OOISCFFL].max_cnt)
- hmc_info->hmc_obj[IRDMA_HMC_IW_OOISCFFL].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_OOISC].cnt /
- hmc_fpm_misc->ooiscf_block_size;
- }
- /**
- * irdma_cfg_fpm_val - configure HMC objects
- * @dev: sc device struct
- * @qp_count: desired qp count
- */
- int irdma_cfg_fpm_val(struct irdma_sc_dev *dev, u32 qp_count)
- {
- struct irdma_virt_mem virt_mem;
- u32 i, mem_size;
- u32 qpwanted, mrwanted, pblewanted;
- u32 powerof2, hte;
- u32 sd_needed;
- u32 sd_diff;
- u32 loop_count = 0;
- struct irdma_hmc_info *hmc_info;
- struct irdma_hmc_fpm_misc *hmc_fpm_misc;
- int ret_code = 0;
- hmc_info = dev->hmc_info;
- hmc_fpm_misc = &dev->hmc_fpm_misc;
- ret_code = irdma_sc_init_iw_hmc(dev, dev->hmc_fn_id);
- if (ret_code) {
- ibdev_dbg(to_ibdev(dev),
- "HMC: irdma_sc_init_iw_hmc returned error_code = %d\n",
- ret_code);
- return ret_code;
- }
- for (i = IRDMA_HMC_IW_QP; i < IRDMA_HMC_IW_MAX; i++)
- hmc_info->hmc_obj[i].cnt = hmc_info->hmc_obj[i].max_cnt;
- sd_needed = irdma_est_sd(dev, hmc_info);
- ibdev_dbg(to_ibdev(dev),
- "HMC: FW max resources sd_needed[%08d] first_sd_index[%04d]\n",
- sd_needed, hmc_info->first_sd_index);
- ibdev_dbg(to_ibdev(dev), "HMC: sd count %d where max sd is %d\n",
- hmc_info->sd_table.sd_cnt, hmc_fpm_misc->max_sds);
- qpwanted = min(qp_count, hmc_info->hmc_obj[IRDMA_HMC_IW_QP].max_cnt);
- powerof2 = 1;
- while (powerof2 <= qpwanted)
- powerof2 *= 2;
- powerof2 /= 2;
- qpwanted = powerof2;
- mrwanted = hmc_info->hmc_obj[IRDMA_HMC_IW_MR].max_cnt;
- pblewanted = hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].max_cnt;
- ibdev_dbg(to_ibdev(dev),
- "HMC: req_qp=%d max_sd=%d, max_qp = %d, max_cq=%d, max_mr=%d, max_pble=%d, mc=%d, av=%d\n",
- qp_count, hmc_fpm_misc->max_sds,
- hmc_info->hmc_obj[IRDMA_HMC_IW_QP].max_cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].max_cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_MR].max_cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].max_cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].max_cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].max_cnt);
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].max_cnt;
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].max_cnt;
- hmc_info->hmc_obj[IRDMA_HMC_IW_ARP].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_ARP].max_cnt;
- hmc_info->hmc_obj[IRDMA_HMC_IW_APBVT_ENTRY].cnt = 1;
- while (irdma_q1_cnt(dev, hmc_info, qpwanted) > hmc_info->hmc_obj[IRDMA_HMC_IW_Q1].max_cnt)
- qpwanted /= 2;
- do {
- ++loop_count;
- hmc_info->hmc_obj[IRDMA_HMC_IW_QP].cnt = qpwanted;
- hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].cnt =
- min(2 * qpwanted, hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].cnt);
- hmc_info->hmc_obj[IRDMA_HMC_IW_RESERVED].cnt = 0; /* Reserved */
- hmc_info->hmc_obj[IRDMA_HMC_IW_MR].cnt = mrwanted;
- hte = round_up(qpwanted + hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].cnt, 512);
- powerof2 = 1;
- while (powerof2 < hte)
- powerof2 *= 2;
- hmc_info->hmc_obj[IRDMA_HMC_IW_HTE].cnt =
- powerof2 * hmc_fpm_misc->ht_multiplier;
- if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1)
- cfg_fpm_value_gen_1(dev, hmc_info, qpwanted);
- else
- cfg_fpm_value_gen_2(dev, hmc_info, qpwanted);
- hmc_info->hmc_obj[IRDMA_HMC_IW_Q1].cnt = irdma_q1_cnt(dev, hmc_info, qpwanted);
- hmc_info->hmc_obj[IRDMA_HMC_IW_XFFL].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_XF].cnt / hmc_fpm_misc->xf_block_size;
- hmc_info->hmc_obj[IRDMA_HMC_IW_Q1FL].cnt =
- hmc_info->hmc_obj[IRDMA_HMC_IW_Q1].cnt / hmc_fpm_misc->q1_block_size;
- hmc_info->hmc_obj[IRDMA_HMC_IW_TIMER].cnt =
- (round_up(qpwanted, 512) / 512 + 1) * hmc_fpm_misc->timer_bucket;
- hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt = pblewanted;
- sd_needed = irdma_est_sd(dev, hmc_info);
- ibdev_dbg(to_ibdev(dev),
- "HMC: sd_needed = %d, hmc_fpm_misc->max_sds=%d, mrwanted=%d, pblewanted=%d qpwanted=%d\n",
- sd_needed, hmc_fpm_misc->max_sds, mrwanted,
- pblewanted, qpwanted);
- /* Do not reduce resources further. All objects fit with max SDs */
- if (sd_needed <= hmc_fpm_misc->max_sds)
- break;
- sd_diff = sd_needed - hmc_fpm_misc->max_sds;
- if (sd_diff > 128) {
- if (!(loop_count % 2) && qpwanted > 128) {
- qpwanted /= 2;
- } else {
- mrwanted /= 2;
- pblewanted /= 2;
- }
- continue;
- }
- if (dev->cqp->hmc_profile != IRDMA_HMC_PROFILE_FAVOR_VF &&
- pblewanted > (512 * FPM_MULTIPLIER * sd_diff)) {
- pblewanted -= 256 * FPM_MULTIPLIER * sd_diff;
- continue;
- } else if (pblewanted > (100 * FPM_MULTIPLIER)) {
- pblewanted -= 10 * FPM_MULTIPLIER;
- } else if (pblewanted > FPM_MULTIPLIER) {
- pblewanted -= FPM_MULTIPLIER;
- } else if (qpwanted <= 128) {
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].cnt > 256)
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].cnt /= 2;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt > 256)
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt /= 2;
- }
- if (mrwanted > FPM_MULTIPLIER)
- mrwanted -= FPM_MULTIPLIER;
- if (!(loop_count % 10) && qpwanted > 128) {
- qpwanted /= 2;
- if (hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt > 256)
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt /= 2;
- }
- } while (loop_count < 2000);
- if (sd_needed > hmc_fpm_misc->max_sds) {
- ibdev_dbg(to_ibdev(dev),
- "HMC: cfg_fpm failed loop_cnt=%d, sd_needed=%d, max sd count %d\n",
- loop_count, sd_needed, hmc_info->sd_table.sd_cnt);
- return -EINVAL;
- }
- if (loop_count > 1 && sd_needed < hmc_fpm_misc->max_sds) {
- pblewanted += (hmc_fpm_misc->max_sds - sd_needed) * 256 *
- FPM_MULTIPLIER;
- hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt = pblewanted;
- sd_needed = irdma_est_sd(dev, hmc_info);
- }
- ibdev_dbg(to_ibdev(dev),
- "HMC: loop_cnt=%d, sd_needed=%d, qpcnt = %d, cqcnt=%d, mrcnt=%d, pblecnt=%d, mc=%d, ah=%d, max sd count %d, first sd index %d\n",
- loop_count, sd_needed,
- hmc_info->hmc_obj[IRDMA_HMC_IW_QP].cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_MR].cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_PBLE].cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIMC].cnt,
- hmc_info->hmc_obj[IRDMA_HMC_IW_FSIAV].cnt,
- hmc_info->sd_table.sd_cnt, hmc_info->first_sd_index);
- ret_code = irdma_sc_cfg_iw_fpm(dev, dev->hmc_fn_id);
- if (ret_code) {
- ibdev_dbg(to_ibdev(dev),
- "HMC: cfg_iw_fpm returned error_code[x%08X]\n",
- readl(dev->hw_regs[IRDMA_CQPERRCODES]));
- return ret_code;
- }
- mem_size = sizeof(struct irdma_hmc_sd_entry) *
- (hmc_info->sd_table.sd_cnt + hmc_info->first_sd_index + 1);
- virt_mem.size = mem_size;
- virt_mem.va = kzalloc(virt_mem.size, GFP_KERNEL);
- if (!virt_mem.va) {
- ibdev_dbg(to_ibdev(dev),
- "HMC: failed to allocate memory for sd_entry buffer\n");
- return -ENOMEM;
- }
- hmc_info->sd_table.sd_entry = virt_mem.va;
- return ret_code;
- }
- /**
- * irdma_exec_cqp_cmd - execute cqp cmd when wqe are available
- * @dev: rdma device
- * @pcmdinfo: cqp command info
- */
- static int irdma_exec_cqp_cmd(struct irdma_sc_dev *dev,
- struct cqp_cmds_info *pcmdinfo)
- {
- int status;
- struct irdma_dma_mem val_mem;
- bool alloc = false;
- dev->cqp_cmd_stats[pcmdinfo->cqp_cmd]++;
- switch (pcmdinfo->cqp_cmd) {
- case IRDMA_OP_CEQ_DESTROY:
- status = irdma_sc_ceq_destroy(pcmdinfo->in.u.ceq_destroy.ceq,
- pcmdinfo->in.u.ceq_destroy.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_AEQ_DESTROY:
- status = irdma_sc_aeq_destroy(pcmdinfo->in.u.aeq_destroy.aeq,
- pcmdinfo->in.u.aeq_destroy.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_CEQ_CREATE:
- status = irdma_sc_ceq_create(pcmdinfo->in.u.ceq_create.ceq,
- pcmdinfo->in.u.ceq_create.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_AEQ_CREATE:
- status = irdma_sc_aeq_create(pcmdinfo->in.u.aeq_create.aeq,
- pcmdinfo->in.u.aeq_create.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_QP_UPLOAD_CONTEXT:
- status = irdma_sc_qp_upload_context(pcmdinfo->in.u.qp_upload_context.dev,
- &pcmdinfo->in.u.qp_upload_context.info,
- pcmdinfo->in.u.qp_upload_context.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_CQ_CREATE:
- status = irdma_sc_cq_create(pcmdinfo->in.u.cq_create.cq,
- pcmdinfo->in.u.cq_create.scratch,
- pcmdinfo->in.u.cq_create.check_overflow,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_CQ_MODIFY:
- status = irdma_sc_cq_modify(pcmdinfo->in.u.cq_modify.cq,
- &pcmdinfo->in.u.cq_modify.info,
- pcmdinfo->in.u.cq_modify.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_CQ_DESTROY:
- status = irdma_sc_cq_destroy(pcmdinfo->in.u.cq_destroy.cq,
- pcmdinfo->in.u.cq_destroy.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_QP_FLUSH_WQES:
- status = irdma_sc_qp_flush_wqes(pcmdinfo->in.u.qp_flush_wqes.qp,
- &pcmdinfo->in.u.qp_flush_wqes.info,
- pcmdinfo->in.u.qp_flush_wqes.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_GEN_AE:
- status = irdma_sc_gen_ae(pcmdinfo->in.u.gen_ae.qp,
- &pcmdinfo->in.u.gen_ae.info,
- pcmdinfo->in.u.gen_ae.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_MANAGE_PUSH_PAGE:
- status = irdma_sc_manage_push_page(pcmdinfo->in.u.manage_push_page.cqp,
- &pcmdinfo->in.u.manage_push_page.info,
- pcmdinfo->in.u.manage_push_page.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_UPDATE_PE_SDS:
- status = irdma_update_pe_sds(pcmdinfo->in.u.update_pe_sds.dev,
- &pcmdinfo->in.u.update_pe_sds.info,
- pcmdinfo->in.u.update_pe_sds.scratch);
- break;
- case IRDMA_OP_MANAGE_HMC_PM_FUNC_TABLE:
- /* switch to calling through the call table */
- status =
- irdma_sc_manage_hmc_pm_func_table(pcmdinfo->in.u.manage_hmc_pm.dev->cqp,
- &pcmdinfo->in.u.manage_hmc_pm.info,
- pcmdinfo->in.u.manage_hmc_pm.scratch,
- true);
- break;
- case IRDMA_OP_SUSPEND:
- status = irdma_sc_suspend_qp(pcmdinfo->in.u.suspend_resume.cqp,
- pcmdinfo->in.u.suspend_resume.qp,
- pcmdinfo->in.u.suspend_resume.scratch);
- break;
- case IRDMA_OP_RESUME:
- status = irdma_sc_resume_qp(pcmdinfo->in.u.suspend_resume.cqp,
- pcmdinfo->in.u.suspend_resume.qp,
- pcmdinfo->in.u.suspend_resume.scratch);
- break;
- case IRDMA_OP_QUERY_FPM_VAL:
- val_mem.pa = pcmdinfo->in.u.query_fpm_val.fpm_val_pa;
- val_mem.va = pcmdinfo->in.u.query_fpm_val.fpm_val_va;
- status = irdma_sc_query_fpm_val(pcmdinfo->in.u.query_fpm_val.cqp,
- pcmdinfo->in.u.query_fpm_val.scratch,
- pcmdinfo->in.u.query_fpm_val.hmc_fn_id,
- &val_mem, true, IRDMA_CQP_WAIT_EVENT);
- break;
- case IRDMA_OP_COMMIT_FPM_VAL:
- val_mem.pa = pcmdinfo->in.u.commit_fpm_val.fpm_val_pa;
- val_mem.va = pcmdinfo->in.u.commit_fpm_val.fpm_val_va;
- status = irdma_sc_commit_fpm_val(pcmdinfo->in.u.commit_fpm_val.cqp,
- pcmdinfo->in.u.commit_fpm_val.scratch,
- pcmdinfo->in.u.commit_fpm_val.hmc_fn_id,
- &val_mem,
- true,
- IRDMA_CQP_WAIT_EVENT);
- break;
- case IRDMA_OP_STATS_ALLOCATE:
- alloc = true;
- fallthrough;
- case IRDMA_OP_STATS_FREE:
- status = irdma_sc_manage_stats_inst(pcmdinfo->in.u.stats_manage.cqp,
- &pcmdinfo->in.u.stats_manage.info,
- alloc,
- pcmdinfo->in.u.stats_manage.scratch);
- break;
- case IRDMA_OP_STATS_GATHER:
- status = irdma_sc_gather_stats(pcmdinfo->in.u.stats_gather.cqp,
- &pcmdinfo->in.u.stats_gather.info,
- pcmdinfo->in.u.stats_gather.scratch);
- break;
- case IRDMA_OP_WS_MODIFY_NODE:
- status = irdma_sc_manage_ws_node(pcmdinfo->in.u.ws_node.cqp,
- &pcmdinfo->in.u.ws_node.info,
- IRDMA_MODIFY_NODE,
- pcmdinfo->in.u.ws_node.scratch);
- break;
- case IRDMA_OP_WS_DELETE_NODE:
- status = irdma_sc_manage_ws_node(pcmdinfo->in.u.ws_node.cqp,
- &pcmdinfo->in.u.ws_node.info,
- IRDMA_DEL_NODE,
- pcmdinfo->in.u.ws_node.scratch);
- break;
- case IRDMA_OP_WS_ADD_NODE:
- status = irdma_sc_manage_ws_node(pcmdinfo->in.u.ws_node.cqp,
- &pcmdinfo->in.u.ws_node.info,
- IRDMA_ADD_NODE,
- pcmdinfo->in.u.ws_node.scratch);
- break;
- case IRDMA_OP_SET_UP_MAP:
- status = irdma_sc_set_up_map(pcmdinfo->in.u.up_map.cqp,
- &pcmdinfo->in.u.up_map.info,
- pcmdinfo->in.u.up_map.scratch);
- break;
- case IRDMA_OP_QUERY_RDMA_FEATURES:
- status = irdma_sc_query_rdma_features(pcmdinfo->in.u.query_rdma.cqp,
- &pcmdinfo->in.u.query_rdma.query_buff_mem,
- pcmdinfo->in.u.query_rdma.scratch);
- break;
- case IRDMA_OP_DELETE_ARP_CACHE_ENTRY:
- status = irdma_sc_del_arp_cache_entry(pcmdinfo->in.u.del_arp_cache_entry.cqp,
- pcmdinfo->in.u.del_arp_cache_entry.scratch,
- pcmdinfo->in.u.del_arp_cache_entry.arp_index,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_MANAGE_APBVT_ENTRY:
- status = irdma_sc_manage_apbvt_entry(pcmdinfo->in.u.manage_apbvt_entry.cqp,
- &pcmdinfo->in.u.manage_apbvt_entry.info,
- pcmdinfo->in.u.manage_apbvt_entry.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_MANAGE_QHASH_TABLE_ENTRY:
- status = irdma_sc_manage_qhash_table_entry(pcmdinfo->in.u.manage_qhash_table_entry.cqp,
- &pcmdinfo->in.u.manage_qhash_table_entry.info,
- pcmdinfo->in.u.manage_qhash_table_entry.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_QP_MODIFY:
- status = irdma_sc_qp_modify(pcmdinfo->in.u.qp_modify.qp,
- &pcmdinfo->in.u.qp_modify.info,
- pcmdinfo->in.u.qp_modify.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_QP_CREATE:
- status = irdma_sc_qp_create(pcmdinfo->in.u.qp_create.qp,
- &pcmdinfo->in.u.qp_create.info,
- pcmdinfo->in.u.qp_create.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_QP_DESTROY:
- status = irdma_sc_qp_destroy(pcmdinfo->in.u.qp_destroy.qp,
- pcmdinfo->in.u.qp_destroy.scratch,
- pcmdinfo->in.u.qp_destroy.remove_hash_idx,
- pcmdinfo->in.u.qp_destroy.ignore_mw_bnd,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_ALLOC_STAG:
- status = irdma_sc_alloc_stag(pcmdinfo->in.u.alloc_stag.dev,
- &pcmdinfo->in.u.alloc_stag.info,
- pcmdinfo->in.u.alloc_stag.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_MR_REG_NON_SHARED:
- status = irdma_sc_mr_reg_non_shared(pcmdinfo->in.u.mr_reg_non_shared.dev,
- &pcmdinfo->in.u.mr_reg_non_shared.info,
- pcmdinfo->in.u.mr_reg_non_shared.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_DEALLOC_STAG:
- status = irdma_sc_dealloc_stag(pcmdinfo->in.u.dealloc_stag.dev,
- &pcmdinfo->in.u.dealloc_stag.info,
- pcmdinfo->in.u.dealloc_stag.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_MW_ALLOC:
- status = irdma_sc_mw_alloc(pcmdinfo->in.u.mw_alloc.dev,
- &pcmdinfo->in.u.mw_alloc.info,
- pcmdinfo->in.u.mw_alloc.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_ADD_ARP_CACHE_ENTRY:
- status = irdma_sc_add_arp_cache_entry(pcmdinfo->in.u.add_arp_cache_entry.cqp,
- &pcmdinfo->in.u.add_arp_cache_entry.info,
- pcmdinfo->in.u.add_arp_cache_entry.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_ALLOC_LOCAL_MAC_ENTRY:
- status = irdma_sc_alloc_local_mac_entry(pcmdinfo->in.u.alloc_local_mac_entry.cqp,
- pcmdinfo->in.u.alloc_local_mac_entry.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_ADD_LOCAL_MAC_ENTRY:
- status = irdma_sc_add_local_mac_entry(pcmdinfo->in.u.add_local_mac_entry.cqp,
- &pcmdinfo->in.u.add_local_mac_entry.info,
- pcmdinfo->in.u.add_local_mac_entry.scratch,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_DELETE_LOCAL_MAC_ENTRY:
- status = irdma_sc_del_local_mac_entry(pcmdinfo->in.u.del_local_mac_entry.cqp,
- pcmdinfo->in.u.del_local_mac_entry.scratch,
- pcmdinfo->in.u.del_local_mac_entry.entry_idx,
- pcmdinfo->in.u.del_local_mac_entry.ignore_ref_count,
- pcmdinfo->post_sq);
- break;
- case IRDMA_OP_AH_CREATE:
- status = irdma_sc_create_ah(pcmdinfo->in.u.ah_create.cqp,
- &pcmdinfo->in.u.ah_create.info,
- pcmdinfo->in.u.ah_create.scratch);
- break;
- case IRDMA_OP_AH_DESTROY:
- status = irdma_sc_destroy_ah(pcmdinfo->in.u.ah_destroy.cqp,
- &pcmdinfo->in.u.ah_destroy.info,
- pcmdinfo->in.u.ah_destroy.scratch);
- break;
- case IRDMA_OP_MC_CREATE:
- status = irdma_sc_create_mcast_grp(pcmdinfo->in.u.mc_create.cqp,
- &pcmdinfo->in.u.mc_create.info,
- pcmdinfo->in.u.mc_create.scratch);
- break;
- case IRDMA_OP_MC_DESTROY:
- status = irdma_sc_destroy_mcast_grp(pcmdinfo->in.u.mc_destroy.cqp,
- &pcmdinfo->in.u.mc_destroy.info,
- pcmdinfo->in.u.mc_destroy.scratch);
- break;
- case IRDMA_OP_MC_MODIFY:
- status = irdma_sc_modify_mcast_grp(pcmdinfo->in.u.mc_modify.cqp,
- &pcmdinfo->in.u.mc_modify.info,
- pcmdinfo->in.u.mc_modify.scratch);
- break;
- default:
- status = -EOPNOTSUPP;
- break;
- }
- return status;
- }
- /**
- * irdma_process_cqp_cmd - process all cqp commands
- * @dev: sc device struct
- * @pcmdinfo: cqp command info
- */
- int irdma_process_cqp_cmd(struct irdma_sc_dev *dev,
- struct cqp_cmds_info *pcmdinfo)
- {
- int status = 0;
- unsigned long flags;
- spin_lock_irqsave(&dev->cqp_lock, flags);
- if (list_empty(&dev->cqp_cmd_head) && !irdma_cqp_ring_full(dev->cqp))
- status = irdma_exec_cqp_cmd(dev, pcmdinfo);
- else
- list_add_tail(&pcmdinfo->cqp_cmd_entry, &dev->cqp_cmd_head);
- spin_unlock_irqrestore(&dev->cqp_lock, flags);
- return status;
- }
- /**
- * irdma_process_bh - called from tasklet for cqp list
- * @dev: sc device struct
- */
- int irdma_process_bh(struct irdma_sc_dev *dev)
- {
- int status = 0;
- struct cqp_cmds_info *pcmdinfo;
- unsigned long flags;
- spin_lock_irqsave(&dev->cqp_lock, flags);
- while (!list_empty(&dev->cqp_cmd_head) &&
- !irdma_cqp_ring_full(dev->cqp)) {
- pcmdinfo = (struct cqp_cmds_info *)irdma_remove_cqp_head(dev);
- status = irdma_exec_cqp_cmd(dev, pcmdinfo);
- if (status)
- break;
- }
- spin_unlock_irqrestore(&dev->cqp_lock, flags);
- return status;
- }
- /**
- * irdma_cfg_aeq- Configure AEQ interrupt
- * @dev: pointer to the device structure
- * @idx: vector index
- * @enable: True to enable, False disables
- */
- void irdma_cfg_aeq(struct irdma_sc_dev *dev, u32 idx, bool enable)
- {
- u32 reg_val;
- reg_val = FIELD_PREP(IRDMA_PFINT_AEQCTL_CAUSE_ENA, enable) |
- FIELD_PREP(IRDMA_PFINT_AEQCTL_MSIX_INDX, idx) |
- FIELD_PREP(IRDMA_PFINT_AEQCTL_ITR_INDX, 3);
- writel(reg_val, dev->hw_regs[IRDMA_PFINT_AEQCTL]);
- }
- /**
- * sc_vsi_update_stats - Update statistics
- * @vsi: sc_vsi instance to update
- */
- void sc_vsi_update_stats(struct irdma_sc_vsi *vsi)
- {
- struct irdma_gather_stats *gather_stats;
- struct irdma_gather_stats *last_gather_stats;
- gather_stats = vsi->pestat->gather_info.gather_stats_va;
- last_gather_stats = vsi->pestat->gather_info.last_gather_stats_va;
- irdma_update_stats(&vsi->pestat->hw_stats, gather_stats,
- last_gather_stats);
- }
- /**
- * irdma_wait_pe_ready - Check if firmware is ready
- * @dev: provides access to registers
- */
- static int irdma_wait_pe_ready(struct irdma_sc_dev *dev)
- {
- u32 statuscpu0;
- u32 statuscpu1;
- u32 statuscpu2;
- u32 retrycount = 0;
- do {
- statuscpu0 = readl(dev->hw_regs[IRDMA_GLPE_CPUSTATUS0]);
- statuscpu1 = readl(dev->hw_regs[IRDMA_GLPE_CPUSTATUS1]);
- statuscpu2 = readl(dev->hw_regs[IRDMA_GLPE_CPUSTATUS2]);
- if (statuscpu0 == 0x80 && statuscpu1 == 0x80 &&
- statuscpu2 == 0x80)
- return 0;
- mdelay(1000);
- } while (retrycount++ < dev->hw_attrs.max_pe_ready_count);
- return -1;
- }
- static inline void irdma_sc_init_hw(struct irdma_sc_dev *dev)
- {
- switch (dev->hw_attrs.uk_attrs.hw_rev) {
- case IRDMA_GEN_1:
- i40iw_init_hw(dev);
- break;
- case IRDMA_GEN_2:
- icrdma_init_hw(dev);
- break;
- }
- }
- /**
- * irdma_sc_dev_init - Initialize control part of device
- * @ver: version
- * @dev: Device pointer
- * @info: Device init info
- */
- int irdma_sc_dev_init(enum irdma_vers ver, struct irdma_sc_dev *dev,
- struct irdma_device_init_info *info)
- {
- u32 val;
- int ret_code = 0;
- u8 db_size;
- INIT_LIST_HEAD(&dev->cqp_cmd_head); /* for CQP command backlog */
- mutex_init(&dev->ws_mutex);
- dev->hmc_fn_id = info->hmc_fn_id;
- dev->fpm_query_buf_pa = info->fpm_query_buf_pa;
- dev->fpm_query_buf = info->fpm_query_buf;
- dev->fpm_commit_buf_pa = info->fpm_commit_buf_pa;
- dev->fpm_commit_buf = info->fpm_commit_buf;
- dev->hw = info->hw;
- dev->hw->hw_addr = info->bar0;
- /* Setup the hardware limits, hmc may limit further */
- dev->hw_attrs.min_hw_qp_id = IRDMA_MIN_IW_QP_ID;
- dev->hw_attrs.min_hw_aeq_size = IRDMA_MIN_AEQ_ENTRIES;
- dev->hw_attrs.max_hw_aeq_size = IRDMA_MAX_AEQ_ENTRIES;
- dev->hw_attrs.min_hw_ceq_size = IRDMA_MIN_CEQ_ENTRIES;
- dev->hw_attrs.max_hw_ceq_size = IRDMA_MAX_CEQ_ENTRIES;
- dev->hw_attrs.uk_attrs.min_hw_cq_size = IRDMA_MIN_CQ_SIZE;
- dev->hw_attrs.uk_attrs.max_hw_cq_size = IRDMA_MAX_CQ_SIZE;
- dev->hw_attrs.uk_attrs.max_hw_wq_frags = IRDMA_MAX_WQ_FRAGMENT_COUNT;
- dev->hw_attrs.uk_attrs.max_hw_read_sges = IRDMA_MAX_SGE_RD;
- dev->hw_attrs.max_hw_outbound_msg_size = IRDMA_MAX_OUTBOUND_MSG_SIZE;
- dev->hw_attrs.max_mr_size = IRDMA_MAX_MR_SIZE;
- dev->hw_attrs.max_hw_inbound_msg_size = IRDMA_MAX_INBOUND_MSG_SIZE;
- dev->hw_attrs.max_hw_device_pages = IRDMA_MAX_PUSH_PAGE_COUNT;
- dev->hw_attrs.uk_attrs.max_hw_inline = IRDMA_MAX_INLINE_DATA_SIZE;
- dev->hw_attrs.max_hw_wqes = IRDMA_MAX_WQ_ENTRIES;
- dev->hw_attrs.max_qp_wr = IRDMA_MAX_QP_WRS(IRDMA_MAX_QUANTA_PER_WR);
- dev->hw_attrs.uk_attrs.max_hw_rq_quanta = IRDMA_QP_SW_MAX_RQ_QUANTA;
- dev->hw_attrs.uk_attrs.max_hw_wq_quanta = IRDMA_QP_SW_MAX_WQ_QUANTA;
- dev->hw_attrs.max_hw_pds = IRDMA_MAX_PDS;
- dev->hw_attrs.max_hw_ena_vf_count = IRDMA_MAX_PE_ENA_VF_COUNT;
- dev->hw_attrs.max_pe_ready_count = 14;
- dev->hw_attrs.max_done_count = IRDMA_DONE_COUNT;
- dev->hw_attrs.max_sleep_count = IRDMA_SLEEP_COUNT;
- dev->hw_attrs.max_cqp_compl_wait_time_ms = CQP_COMPL_WAIT_TIME_MS;
- dev->hw_attrs.uk_attrs.hw_rev = ver;
- irdma_sc_init_hw(dev);
- if (irdma_wait_pe_ready(dev))
- return -ETIMEDOUT;
- val = readl(dev->hw_regs[IRDMA_GLPCI_LBARCTRL]);
- db_size = (u8)FIELD_GET(IRDMA_GLPCI_LBARCTRL_PE_DB_SIZE, val);
- if (db_size != IRDMA_PE_DB_SIZE_4M && db_size != IRDMA_PE_DB_SIZE_8M) {
- ibdev_dbg(to_ibdev(dev),
- "DEV: RDMA PE doorbell is not enabled in CSR val 0x%x db_size=%d\n",
- val, db_size);
- return -ENODEV;
- }
- dev->db_addr = dev->hw->hw_addr + (uintptr_t)dev->hw_regs[IRDMA_DB_ADDR_OFFSET];
- return ret_code;
- }
- /**
- * irdma_update_stats - Update statistics
- * @hw_stats: hw_stats instance to update
- * @gather_stats: updated stat counters
- * @last_gather_stats: last stat counters
- */
- void irdma_update_stats(struct irdma_dev_hw_stats *hw_stats,
- struct irdma_gather_stats *gather_stats,
- struct irdma_gather_stats *last_gather_stats)
- {
- u64 *stats_val = hw_stats->stats_val_32;
- stats_val[IRDMA_HW_STAT_INDEX_RXVLANERR] +=
- IRDMA_STATS_DELTA(gather_stats->rxvlanerr,
- last_gather_stats->rxvlanerr,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXDISCARD] +=
- IRDMA_STATS_DELTA(gather_stats->ip4rxdiscard,
- last_gather_stats->ip4rxdiscard,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXTRUNC] +=
- IRDMA_STATS_DELTA(gather_stats->ip4rxtrunc,
- last_gather_stats->ip4rxtrunc,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP4TXNOROUTE] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txnoroute,
- last_gather_stats->ip4txnoroute,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXDISCARD] +=
- IRDMA_STATS_DELTA(gather_stats->ip6rxdiscard,
- last_gather_stats->ip6rxdiscard,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXTRUNC] +=
- IRDMA_STATS_DELTA(gather_stats->ip6rxtrunc,
- last_gather_stats->ip6rxtrunc,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_IP6TXNOROUTE] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txnoroute,
- last_gather_stats->ip6txnoroute,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_TCPRTXSEG] +=
- IRDMA_STATS_DELTA(gather_stats->tcprtxseg,
- last_gather_stats->tcprtxseg,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_TCPRXOPTERR] +=
- IRDMA_STATS_DELTA(gather_stats->tcprxopterr,
- last_gather_stats->tcprxopterr,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_TCPRXPROTOERR] +=
- IRDMA_STATS_DELTA(gather_stats->tcprxprotoerr,
- last_gather_stats->tcprxprotoerr,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_RXRPCNPHANDLED] +=
- IRDMA_STATS_DELTA(gather_stats->rxrpcnphandled,
- last_gather_stats->rxrpcnphandled,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_RXRPCNPIGNORED] +=
- IRDMA_STATS_DELTA(gather_stats->rxrpcnpignored,
- last_gather_stats->rxrpcnpignored,
- IRDMA_MAX_STATS_32);
- stats_val[IRDMA_HW_STAT_INDEX_TXNPCNPSENT] +=
- IRDMA_STATS_DELTA(gather_stats->txnpcnpsent,
- last_gather_stats->txnpcnpsent,
- IRDMA_MAX_STATS_32);
- stats_val = hw_stats->stats_val_64;
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXOCTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4rxocts,
- last_gather_stats->ip4rxocts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4rxpkts,
- last_gather_stats->ip4rxpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXFRAGS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txfrag,
- last_gather_stats->ip4txfrag,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4RXMCPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4rxmcpkts,
- last_gather_stats->ip4rxmcpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4TXOCTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txocts,
- last_gather_stats->ip4txocts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4TXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txpkts,
- last_gather_stats->ip4txpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4TXFRAGS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txfrag,
- last_gather_stats->ip4txfrag,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP4TXMCPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip4txmcpkts,
- last_gather_stats->ip4txmcpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXOCTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6rxocts,
- last_gather_stats->ip6rxocts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6rxpkts,
- last_gather_stats->ip6rxpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXFRAGS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txfrags,
- last_gather_stats->ip6txfrags,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6RXMCPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6rxmcpkts,
- last_gather_stats->ip6rxmcpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6TXOCTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txocts,
- last_gather_stats->ip6txocts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6TXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txpkts,
- last_gather_stats->ip6txpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6TXFRAGS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txfrags,
- last_gather_stats->ip6txfrags,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_IP6TXMCPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->ip6txmcpkts,
- last_gather_stats->ip6txmcpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_TCPRXSEGS] +=
- IRDMA_STATS_DELTA(gather_stats->tcprxsegs,
- last_gather_stats->tcprxsegs,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_TCPTXSEG] +=
- IRDMA_STATS_DELTA(gather_stats->tcptxsegs,
- last_gather_stats->tcptxsegs,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMARXRDS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmarxrds,
- last_gather_stats->rdmarxrds,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMARXSNDS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmarxsnds,
- last_gather_stats->rdmarxsnds,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMARXWRS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmarxwrs,
- last_gather_stats->rdmarxwrs,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMATXRDS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmatxrds,
- last_gather_stats->rdmatxrds,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMATXSNDS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmatxsnds,
- last_gather_stats->rdmatxsnds,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMATXWRS] +=
- IRDMA_STATS_DELTA(gather_stats->rdmatxwrs,
- last_gather_stats->rdmatxwrs,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMAVBND] +=
- IRDMA_STATS_DELTA(gather_stats->rdmavbn,
- last_gather_stats->rdmavbn,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RDMAVINV] +=
- IRDMA_STATS_DELTA(gather_stats->rdmavinv,
- last_gather_stats->rdmavinv,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_UDPRXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->udprxpkts,
- last_gather_stats->udprxpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_UDPTXPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->udptxpkts,
- last_gather_stats->udptxpkts,
- IRDMA_MAX_STATS_48);
- stats_val[IRDMA_HW_STAT_INDEX_RXNPECNMARKEDPKTS] +=
- IRDMA_STATS_DELTA(gather_stats->rxnpecnmrkpkts,
- last_gather_stats->rxnpecnmrkpkts,
- IRDMA_MAX_STATS_48);
- memcpy(last_gather_stats, gather_stats, sizeof(*last_gather_stats));
- }
|